commit stringlengths 40 40 | old_file stringlengths 4 106 | new_file stringlengths 4 106 | old_contents stringlengths 10 2.94k | new_contents stringlengths 21 2.95k | subject stringlengths 16 444 | message stringlengths 17 2.63k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 7 43k | ndiff stringlengths 52 3.31k | instruction stringlengths 16 444 | content stringlengths 133 4.32k | diff stringlengths 49 3.61k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4d19cc36e866c8e21a526cd228f170ffd177292b | run_ctest.py | run_ctest.py |
import platform
import os
import subprocess
import sys
if platform.system() == "Windows":
import distutils.msvc9compiler as msvc
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if len(sys.argv) > 1:
msvc.find_vcvarsall = lambda _: sys.argv[1]
envs = msvc.query_vcvarsall(sys.argv[2])
for k,v in envs.items():
k = k.upper()
v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";"))
v = v.replace("'\''",r"'\'\\\'\''")
print "export %(k)s='\''%(v)s'\''" % locals()
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.")
|
import platform
import os
import subprocess
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.")
| Test travisci windows build workaround. (3) | Test travisci windows build workaround. (3)
| Python | unlicense | h-s-c/ci-tools |
import platform
import os
import subprocess
- import sys
-
- if platform.system() == "Windows":
- import distutils.msvc9compiler as msvc
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
- if len(sys.argv) > 1:
- msvc.find_vcvarsall = lambda _: sys.argv[1]
- envs = msvc.query_vcvarsall(sys.argv[2])
- for k,v in envs.items():
- k = k.upper()
- v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";"))
- v = v.replace("'\''",r"'\'\\\'\''")
- print "export %(k)s='\''%(v)s'\''" % locals()
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.")
| Test travisci windows build workaround. (3) | ## Code Before:
import platform
import os
import subprocess
import sys
if platform.system() == "Windows":
import distutils.msvc9compiler as msvc
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if len(sys.argv) > 1:
msvc.find_vcvarsall = lambda _: sys.argv[1]
envs = msvc.query_vcvarsall(sys.argv[2])
for k,v in envs.items():
k = k.upper()
v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";"))
v = v.replace("'\''",r"'\'\\\'\''")
print "export %(k)s='\''%(v)s'\''" % locals()
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.")
## Instruction:
Test travisci windows build workaround. (3)
## Code After:
import platform
import os
import subprocess
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.")
|
import platform
import os
import subprocess
- import sys
-
- if platform.system() == "Windows":
- import distutils.msvc9compiler as msvc
if __name__ == "__main__":
CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools")
CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake")
if platform.system() == "Linux":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
elif platform.system() == "Windows":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
- if len(sys.argv) > 1:
- msvc.find_vcvarsall = lambda _: sys.argv[1]
- envs = msvc.query_vcvarsall(sys.argv[2])
- for k,v in envs.items():
- k = k.upper()
- v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";"))
- v = v.replace("'\''",r"'\'\\\'\''")
- print "export %(k)s='\''%(v)s'\''" % locals()
elif platform.system() == "Darwin":
os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin"))
if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0:
raise Exception("CTest returned an error.") |
b1bf5dfa91f1f7b84512f72d6e5e18c2109f3239 | addic7ed/__init__.py | addic7ed/__init__.py | from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
print(ep.rename(filename), end="\n\n")
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
filename, "green"))
except Exception as e:
print(colored(e, "red"),
end="\n\n")
| from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
print(ep.rename(filename))
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
filename, "green"), end="\n\n")
except Exception as e:
print(colored(e, "red"),
end="\n\n")
| Fix newline output of downloaded srt | Fix newline output of downloaded srt
| Python | mit | Jesus-21/addic7ed | from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
- print(ep.rename(filename), end="\n\n")
+ print(ep.rename(filename))
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
- filename, "green"))
+ filename, "green"), end="\n\n")
except Exception as e:
print(colored(e, "red"),
end="\n\n")
| Fix newline output of downloaded srt | ## Code Before:
from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
print(ep.rename(filename), end="\n\n")
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
filename, "green"))
except Exception as e:
print(colored(e, "red"),
end="\n\n")
## Instruction:
Fix newline output of downloaded srt
## Code After:
from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
print(ep.rename(filename))
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
filename, "green"), end="\n\n")
except Exception as e:
print(colored(e, "red"),
end="\n\n")
| from termcolor import colored
from .parser import Addic7edParser
from .file_crawler import FileCrawler
from .logger import init_logger
from .config import Config
def addic7ed():
try:
init_logger()
Config.load()
main()
except (EOFError, KeyboardInterrupt, SystemExit):
print(colored("\nBye!", "yellow"))
exit(0)
def main():
crawler = FileCrawler()
parser = Addic7edParser()
for filename, ep in crawler.episodes.items():
subs = parser.parse(**ep.infos)
print(ep)
if not subs:
print(colored("No subtitles for %s" % filename, "red"), end="\n\n")
continue
for i, sub in enumerate(subs):
print("[%s] %s" % (colored(i, "yellow"), sub))
if Config.dry_run:
print()
continue
else:
version = input('Download number? ')
if not version:
print(colored("Nothing to do!", "yellow"),
end="\n\n")
continue
try:
if Config.rename != "sub":
filename = subs[int(version)].download()
if filename and Config.rename == "video":
- print(ep.rename(filename), end="\n\n")
? ------------
+ print(ep.rename(filename))
else:
filename = subs[int(version)].download("%s.srt" %
ep.filename)
print(colored("Downloaded %s subtitle file" %
- filename, "green"))
+ filename, "green"), end="\n\n")
? ++++++++++++
except Exception as e:
print(colored(e, "red"),
end="\n\n") |
e790e47e6b87bc2e49e8b74d491eb023c4468254 | src/sentry/web/frontend/csrf_failure.py | src/sentry/web/frontend/csrf_failure.py | from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
from sentry.web.frontend.base import BaseView
class CsrfFailureView(BaseView):
auth_required = False
sudo_required = False
def handle(self, request, reason=""):
context = {
'no_referer': reason == REASON_NO_REFERER
}
return self.respond('sentry/403-csrf-failure.html', status=403)
view = CsrfFailureView.as_view()
| from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.utils.decorators import method_decorator
from sentry.web.helpers import render_to_response
class CsrfFailureView(View):
@method_decorator(csrf_exempt)
def dispatch(self, request, reason=""):
context = {
'no_referer': reason == REASON_NO_REFERER,
'request': request,
}
return render_to_response('sentry/403-csrf-failure.html', context, request,
status=403)
view = CsrfFailureView.as_view()
| Kill possible recursion on csrf decorator | Kill possible recursion on csrf decorator
| Python | bsd-3-clause | boneyao/sentry,jean/sentry,boneyao/sentry,mvaled/sentry,felixbuenemann/sentry,kevinlondon/sentry,TedaLIEz/sentry,JamesMura/sentry,kevinastone/sentry,korealerts1/sentry,JackDanger/sentry,songyi199111/sentry,songyi199111/sentry,fuziontech/sentry,JamesMura/sentry,BuildingLink/sentry,camilonova/sentry,wujuguang/sentry,argonemyth/sentry,wujuguang/sentry,pauloschilling/sentry,zenefits/sentry,nicholasserra/sentry,beeftornado/sentry,ewdurbin/sentry,gg7/sentry,Natim/sentry,vperron/sentry,Natim/sentry,korealerts1/sentry,kevinlondon/sentry,alexm92/sentry,wong2/sentry,gencer/sentry,BayanGroup/sentry,fuziontech/sentry,jean/sentry,JTCunning/sentry,alexm92/sentry,drcapulet/sentry,gencer/sentry,ifduyue/sentry,ewdurbin/sentry,imankulov/sentry,felixbuenemann/sentry,hongliang5623/sentry,wujuguang/sentry,pauloschilling/sentry,drcapulet/sentry,looker/sentry,nicholasserra/sentry,Kryz/sentry,mvaled/sentry,ewdurbin/sentry,wong2/sentry,imankulov/sentry,jean/sentry,kevinastone/sentry,1tush/sentry,mvaled/sentry,fotinakis/sentry,1tush/sentry,gencer/sentry,hongliang5623/sentry,vperron/sentry,looker/sentry,JackDanger/sentry,hongliang5623/sentry,zenefits/sentry,nicholasserra/sentry,zenefits/sentry,jokey2k/sentry,JamesMura/sentry,songyi199111/sentry,1tush/sentry,drcapulet/sentry,ngonzalvez/sentry,mvaled/sentry,Kryz/sentry,BayanGroup/sentry,kevinlondon/sentry,daevaorn/sentry,ifduyue/sentry,ifduyue/sentry,fotinakis/sentry,zenefits/sentry,BuildingLink/sentry,camilonova/sentry,JTCunning/sentry,TedaLIEz/sentry,Natim/sentry,ngonzalvez/sentry,llonchj/sentry,BuildingLink/sentry,argonemyth/sentry,jean/sentry,argonemyth/sentry,wong2/sentry,looker/sentry,looker/sentry,vperron/sentry,TedaLIEz/sentry,ngonzalvez/sentry,JamesMura/sentry,camilonova/sentry,beeftornado/sentry,looker/sentry,jokey2k/sentry,zenefits/sentry,BayanGroup/sentry,JackDanger/sentry,daevaorn/sentry,BuildingLink/sentry,gg7/sentry,BuildingLink/sentry,Kryz/sentry,gg7/sentry,JTCunning/sentry,boneyao/sentry,mitsuhiko/sentry,ifduyue/sentry,llonchj/sentry,mvaled/sentry,fotinakis/sentry,korealerts1/sentry,jean/sentry,mvaled/sentry,llonchj/sentry,mitsuhiko/sentry,felixbuenemann/sentry,beeftornado/sentry,fuziontech/sentry,JamesMura/sentry,kevinastone/sentry,imankulov/sentry,daevaorn/sentry,jokey2k/sentry,gencer/sentry,pauloschilling/sentry,gencer/sentry,ifduyue/sentry,fotinakis/sentry,alexm92/sentry,daevaorn/sentry | from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
+ from django.views.decorators.csrf import csrf_exempt
+ from django.views.generic import View
+ from django.utils.decorators import method_decorator
- from sentry.web.frontend.base import BaseView
+ from sentry.web.helpers import render_to_response
- class CsrfFailureView(BaseView):
+ class CsrfFailureView(View):
+ @method_decorator(csrf_exempt)
- auth_required = False
- sudo_required = False
-
- def handle(self, request, reason=""):
+ def dispatch(self, request, reason=""):
context = {
- 'no_referer': reason == REASON_NO_REFERER
+ 'no_referer': reason == REASON_NO_REFERER,
+ 'request': request,
}
- return self.respond('sentry/403-csrf-failure.html', status=403)
+ return render_to_response('sentry/403-csrf-failure.html', context, request,
+ status=403)
view = CsrfFailureView.as_view()
| Kill possible recursion on csrf decorator | ## Code Before:
from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
from sentry.web.frontend.base import BaseView
class CsrfFailureView(BaseView):
auth_required = False
sudo_required = False
def handle(self, request, reason=""):
context = {
'no_referer': reason == REASON_NO_REFERER
}
return self.respond('sentry/403-csrf-failure.html', status=403)
view = CsrfFailureView.as_view()
## Instruction:
Kill possible recursion on csrf decorator
## Code After:
from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
from django.views.decorators.csrf import csrf_exempt
from django.views.generic import View
from django.utils.decorators import method_decorator
from sentry.web.helpers import render_to_response
class CsrfFailureView(View):
@method_decorator(csrf_exempt)
def dispatch(self, request, reason=""):
context = {
'no_referer': reason == REASON_NO_REFERER,
'request': request,
}
return render_to_response('sentry/403-csrf-failure.html', context, request,
status=403)
view = CsrfFailureView.as_view()
| from __future__ import absolute_import
from django.middleware.csrf import REASON_NO_REFERER
+ from django.views.decorators.csrf import csrf_exempt
+ from django.views.generic import View
+ from django.utils.decorators import method_decorator
- from sentry.web.frontend.base import BaseView
+ from sentry.web.helpers import render_to_response
- class CsrfFailureView(BaseView):
? ----
+ class CsrfFailureView(View):
+ @method_decorator(csrf_exempt)
- auth_required = False
- sudo_required = False
-
- def handle(self, request, reason=""):
? -----
+ def dispatch(self, request, reason=""):
? +++++++
context = {
- 'no_referer': reason == REASON_NO_REFERER
+ 'no_referer': reason == REASON_NO_REFERER,
? +
+ 'request': request,
}
- return self.respond('sentry/403-csrf-failure.html', status=403)
+ return render_to_response('sentry/403-csrf-failure.html', context, request,
+ status=403)
view = CsrfFailureView.as_view() |
33f4036825c6ff4d9df0038471727648e0df100d | feder/virus_scan/engine/base.py | feder/virus_scan/engine/base.py | from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
self.signer.sign(self.name),
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
)
| import urllib.parse
from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
urllib.parse.quote(self.signer.sign(self.name)),
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
)
| Fix urlencode in webhook url | Fix urlencode in webhook url | Python | mit | watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder | + import urllib.parse
+
from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
- self.signer.sign(self.name),
+ urllib.parse.quote(self.signer.sign(self.name)),
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
)
| Fix urlencode in webhook url | ## Code Before:
from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
self.signer.sign(self.name),
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
)
## Instruction:
Fix urlencode in webhook url
## Code After:
import urllib.parse
from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
urllib.parse.quote(self.signer.sign(self.name)),
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
)
| + import urllib.parse
+
from django.contrib.sites.shortcuts import get_current_site
from django.urls import reverse
from django.core.signing import TimestampSigner
class BaseEngine:
def __init__(self):
self.signer = TimestampSigner()
def get_webhook_url(self):
return "{}://{}{}?token={}".format(
"https",
get_current_site(None).domain,
reverse("virus_scan:webhook"),
- self.signer.sign(self.name),
+ urllib.parse.quote(self.signer.sign(self.name)),
? +++++++++++++++++++ +
)
def send_scan(self, this_file, filename):
raise NotImplementedError(
"Provide 'send' in {name}".format(name=self.__class__.__name__)
)
def receive_scan(self, engine_id):
raise NotImplementedError(
"Provide 'receive_scan' in {name}".format(name=self.__class__.__name__)
) |
c621bc7c94dbbeb5540b2ce46437ee24ecbc33dd | test/test_interface.py | test/test_interface.py | from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
| import cloudbridge
from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
def test_library_version(self):
"""
Check that the library version can be retrieved.
"""
self.assertIsNotNone(cloudbridge.get_version(),
"Did not get library version.")
| Add a library version test | Add a library version test
| Python | mit | gvlproject/cloudbridge,ms-azure-cloudbroker/cloudbridge,gvlproject/libcloudbridge | + import cloudbridge
from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
+ def test_library_version(self):
+ """
+ Check that the library version can be retrieved.
+ """
+ self.assertIsNotNone(cloudbridge.get_version(),
+ "Did not get library version.")
+ | Add a library version test | ## Code Before:
from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
## Instruction:
Add a library version test
## Code After:
import cloudbridge
from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
def test_library_version(self):
"""
Check that the library version can be retrieved.
"""
self.assertIsNotNone(cloudbridge.get_version(),
"Did not get library version.")
| + import cloudbridge
from cloudbridge.cloud import interfaces
from test.helpers import ProviderTestBase
class CloudInterfaceTestCase(ProviderTestBase):
def __init__(self, methodName, provider):
super(CloudInterfaceTestCase, self).__init__(
methodName=methodName, provider=provider)
def test_name_property(self):
"""
Name should always return a value and should not raise an exception
"""
assert self.provider.name
def test_has_service_valid_service_type(self):
"""
has_service with a valid service type should return
a boolean and raise no exceptions
"""
for key, value in interfaces.CloudServiceType.__dict__.items():
if not key.startswith("__"):
self.provider.has_service(value)
def test_has_service_invalid_service_type(self):
"""
has_service with an invalid service type should return False
"""
self.assertFalse(
self.provider.has_service("NON_EXISTENT_SERVICE"),
"has_service should not return True for a non-existent service")
+
+ def test_library_version(self):
+ """
+ Check that the library version can be retrieved.
+ """
+ self.assertIsNotNone(cloudbridge.get_version(),
+ "Did not get library version.") |
50f8efd7bcbf032fd0295c460b98640d0bf6c1ed | smithers/smithers/conf/server.py | smithers/smithers/conf/server.py | from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
STATSD_HOST = 'graphite1.private.phx1.mozilla.com'
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500
| from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com')
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500
| Make statsd host configurable via env. | Make statsd host configurable via env.
| Python | mpl-2.0 | mozilla/mrburns,mozilla/mrburns,mozilla/mrburns | from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
- STATSD_HOST = 'graphite1.private.phx1.mozilla.com'
+ STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com')
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500
| Make statsd host configurable via env. | ## Code Before:
from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
STATSD_HOST = 'graphite1.private.phx1.mozilla.com'
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500
## Instruction:
Make statsd host configurable via env.
## Code After:
from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com')
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500
| from os import getenv
GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb'
- STATSD_HOST = 'graphite1.private.phx1.mozilla.com'
+ STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com')
? ++++++++++++++++++++++ +
STATSD_PORT = 8125
STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV'))
COUNTRY_MIN_SHARE = 500 |
1b95969110f97af397cb3314b59c30679911da48 | scripts/scrape-cdc-state-case-counts.py | scripts/scrape-cdc-state-case-counts.py | import requests
import lxml.html
import pandas as pd
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
cells = [ [ td.text_content().strip()
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8")
| import requests
import lxml.html
import pandas as pd
import re
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
paren_pat = re.compile(r"\([^\)]+\)")
def parse_cell(text):
return re.sub(paren_pat, "", text).strip()
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
cells = [ [ parse_cell(td.text_content())
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8")
| Update CDC scraper to handle new format | Update CDC scraper to handle new format
| Python | mit | BuzzFeedNews/zika-data | import requests
import lxml.html
import pandas as pd
+ import re
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
+
+ paren_pat = re.compile(r"\([^\)]+\)")
+
+ def parse_cell(text):
+ return re.sub(paren_pat, "", text).strip()
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
- cells = [ [ td.text_content().strip()
+ cells = [ [ parse_cell(td.text_content())
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8")
| Update CDC scraper to handle new format | ## Code Before:
import requests
import lxml.html
import pandas as pd
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
cells = [ [ td.text_content().strip()
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8")
## Instruction:
Update CDC scraper to handle new format
## Code After:
import requests
import lxml.html
import pandas as pd
import re
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
paren_pat = re.compile(r"\([^\)]+\)")
def parse_cell(text):
return re.sub(paren_pat, "", text).strip()
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
cells = [ [ parse_cell(td.text_content())
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8")
| import requests
import lxml.html
import pandas as pd
+ import re
import sys
URL = "http://www.cdc.gov/zika/geo/united-states.html"
INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ]
COLS = [ "state_or_territory" ] + INT_COLS
+
+ paren_pat = re.compile(r"\([^\)]+\)")
+
+ def parse_cell(text):
+ return re.sub(paren_pat, "", text).strip()
def scrape():
html = requests.get(URL).content
dom = lxml.html.fromstring(html)
table = dom.cssselect("table")[0]
rows = table.cssselect("tr")
- cells = [ [ td.text_content().strip()
? ------- -
+ cells = [ [ parse_cell(td.text_content())
? +++++++++++
for td in tr.cssselect("td") ]
for tr in rows ]
data = [ c for c in cells
if sum(len(x) != 0 for x in c) == 3 ]
df = pd.DataFrame(data, columns=COLS)
df[INT_COLS] = df[INT_COLS].astype(int)
return df
if __name__ == "__main__":
df = scrape()
df.to_csv(sys.stdout, index=False, encoding="utf-8") |
524d5427d54342f26008a5b527140d4158f70edf | tests/test_extension.py | tests/test_extension.py | from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info
| from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
get_websocket().data = None # make sure it's cleared
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info
| Clear websocket data to try and fix Travis | Clear websocket data to try and fix Travis
| Python | agpl-3.0 | palfrey/mopidy-tachikoma,palfrey/mopidy-tachikoma | from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
+ get_websocket().data = None # make sure it's cleared
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info
| Clear websocket data to try and fix Travis | ## Code Before:
from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info
## Instruction:
Clear websocket data to try and fix Travis
## Code After:
from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
get_websocket().data = None # make sure it's cleared
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info
| from __future__ import unicode_literals
import json
from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot
from mopidy_tachikoma import Extension
def test_get_default_config():
ext = Extension()
config = ext.get_default_config()
assert '[tachikoma]' in config
assert 'enabled = true' in config
assert 'slack_token = ' in config
def test_get_config_schema():
ext = Extension()
schema = ext.get_config_schema()
assert 'slack_token' in schema
@patched_bot
def test_can_connect():
make_frontend()
@patched_bot
def test_gets_events():
frontend = make_frontend()
frontend.doSlackLoop(
None, MockTrack(),
[{"type": "message", "channel": "mock_channel"}])
data = json.loads(get_websocket().data)
assert {
'channel': 'mock_channel',
'text': 'Now playing *foo* from *bar*',
'type': 'message'} == data
@patched_bot
def test_says_one_thing_per_channel():
frontend = make_frontend()
song = MockTrack()
+ get_websocket().data = None # make sure it's cleared
frontend.doSlackLoop(
song, song, [{"type": "message", "channel": "mock_channel"}])
assert get_websocket().data is None # same song, no info |
64d7ca9695eed6112c793fda3f2e7fea3751c3cc | tasks.py | tasks.py | from invoke import run
from invoke import task
@task
def clean(docs=False, bytecode=True, extra=''):
patterns = ['build']
if docs:
patterns.append('docs/_build')
if bytecode:
patterns.append('**/*.pyc')
if extra:
patterns.append(extra)
for pattern in patterns:
run("rm -rf %s" % pattern)
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8")
| from invoke import run
from invoke import task
@task
def clean(all=False):
if all:
flag = "--all"
else:
flag = ""
run("python setup.py clean {}".format(flag))
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8")
| Change clean task to use setup.py | Change clean task to use setup.py
| Python | bsd-3-clause | pando85/django-registration,allo-/django-registration,sergafts/django-registration,pando85/django-registration,allo-/django-registration,sergafts/django-registration | from invoke import run
from invoke import task
@task
+ def clean(all=False):
- def clean(docs=False, bytecode=True, extra=''):
- patterns = ['build']
- if docs:
- patterns.append('docs/_build')
- if bytecode:
- patterns.append('**/*.pyc')
- if extra:
+ if all:
- patterns.append(extra)
- for pattern in patterns:
- run("rm -rf %s" % pattern)
+ flag = "--all"
+ else:
+ flag = ""
+ run("python setup.py clean {}".format(flag))
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8")
| Change clean task to use setup.py | ## Code Before:
from invoke import run
from invoke import task
@task
def clean(docs=False, bytecode=True, extra=''):
patterns = ['build']
if docs:
patterns.append('docs/_build')
if bytecode:
patterns.append('**/*.pyc')
if extra:
patterns.append(extra)
for pattern in patterns:
run("rm -rf %s" % pattern)
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8")
## Instruction:
Change clean task to use setup.py
## Code After:
from invoke import run
from invoke import task
@task
def clean(all=False):
if all:
flag = "--all"
else:
flag = ""
run("python setup.py clean {}".format(flag))
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8")
| from invoke import run
from invoke import task
@task
+ def clean(all=False):
- def clean(docs=False, bytecode=True, extra=''):
- patterns = ['build']
- if docs:
- patterns.append('docs/_build')
- if bytecode:
- patterns.append('**/*.pyc')
- if extra:
? ----
+ if all:
? ++
- patterns.append(extra)
- for pattern in patterns:
- run("rm -rf %s" % pattern)
+ flag = "--all"
+ else:
+ flag = ""
+ run("python setup.py clean {}".format(flag))
@task
def build(docs=False):
run("python setup.py build")
if docs:
run("sphinx-build docs docs/_build")
@task
def test():
run("python setup.py test")
@task
def lint():
run("flake8") |
f794c6ed1f6be231d79ac35759ad76270c3e14e0 | brains/mapping/admin.py | brains/mapping/admin.py | from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
('reported_by', 'origin', 'reported_date')
)}
),)
readonly_fields = ['players', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin)
| from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
('reported_by', 'origin'),
'reported_date',
)}
),)
readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined',
'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level',
'players', 'reported_by', 'origin', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin)
| Set everything on the report read only. | Set everything on the report read only.
| Python | bsd-3-clause | crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains | from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
- ('reported_by', 'origin', 'reported_date')
+ ('reported_by', 'origin'),
+ 'reported_date',
)}
),)
- readonly_fields = ['players', 'reported_date']
+
+ readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined',
+ 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level',
+ 'players', 'reported_by', 'origin', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin)
| Set everything on the report read only. | ## Code Before:
from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
('reported_by', 'origin', 'reported_date')
)}
),)
readonly_fields = ['players', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin)
## Instruction:
Set everything on the report read only.
## Code After:
from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
('reported_by', 'origin'),
'reported_date',
)}
),)
readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined',
'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level',
'players', 'reported_by', 'origin', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin)
| from django.contrib import admin
from mapping.models import Location, Report
class LocationAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': (
('name', 'suburb'),
('x', 'y'),
'building_type'
)}
),)
list_display = ['name', 'x', 'y', 'suburb']
list_filter = ['suburb']
search_fields = ['name']
readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb']
actions = None
def has_add_permission(self, request):
return False
class ReportAdmin(admin.ModelAdmin):
fieldsets = ((None,
{'fields': ('location',
('zombies_only', 'inside'),
('is_ruined', 'is_illuminated', 'has_tree'),
('zombies_present', 'barricade_level'),
'players',
- ('reported_by', 'origin', 'reported_date')
? -----------------
+ ('reported_by', 'origin'),
? +
+ 'reported_date',
)}
),)
- readonly_fields = ['players', 'reported_date']
+
+ readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined',
+ 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level',
+ 'players', 'reported_by', 'origin', 'reported_date']
admin.site.register(Location, LocationAdmin)
admin.site.register(Report, ReportAdmin) |
7a97694d14e9ed02ae38ab713267d4c722079c9c | kolibri/core/webpack/test/test_webpack_tags.py | kolibri/core/webpack/test/test_webpack_tags.py | from __future__ import absolute_import, print_function, unicode_literals
from django.test import TestCase
from ..hooks import WebpackBundleHook
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
entry_file = "assets/src/kolibri_core_app.js"
class KolibriTagNavigationTestCase(TestCase):
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
TestHook().render_to_html()
)
| from __future__ import absolute_import, print_function, unicode_literals
import json
import tempfile
from django.test import TestCase
from ..hooks import WebpackBundleHook
TEST_STATS_FILE = None
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
src_file = "assets/src/kolibri_core_app.js"
@property
def stats_file(self):
return TEST_STATS_FILE.name
class KolibriTagNavigationTestCase(TestCase):
def setUp(self):
global TEST_STATS_FILE
TestCase.setUp(self)
TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False)
self.test_hook = TestHook()
json.dump(
{
"status": "done",
"chunks": {
"non_default_frontend": [
{
"name": "non_default_frontend-2c4fb3d6a29238b06f84.js",
"publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js",
"path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js"
}
]
},
"publicPath": "default_frontend/"
},
TEST_STATS_FILE
)
TEST_STATS_FILE.close()
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
self.test_hook.render_to_html()
)
| Create a faked hook that does not try to read a live stats file | Create a faked hook that does not try to read a live stats file
| Python | mit | jonboiser/kolibri,jtamiace/kolibri,66eli77/kolibri,66eli77/kolibri,mrpau/kolibri,rtibbles/kolibri,learningequality/kolibri,lyw07/kolibri,jtamiace/kolibri,jtamiace/kolibri,mrpau/kolibri,ralphiee22/kolibri,jayoshih/kolibri,learningequality/kolibri,DXCanas/kolibri,jonboiser/kolibri,benjaoming/kolibri,christianmemije/kolibri,indirectlylit/kolibri,benjaoming/kolibri,lyw07/kolibri,aronasorman/kolibri,66eli77/kolibri,MCGallaspy/kolibri,jayoshih/kolibri,DXCanas/kolibri,jonboiser/kolibri,MCGallaspy/kolibri,aronasorman/kolibri,whitzhu/kolibri,DXCanas/kolibri,whitzhu/kolibri,benjaoming/kolibri,rtibbles/kolibri,mrpau/kolibri,lyw07/kolibri,learningequality/kolibri,jamalex/kolibri,rtibbles/kolibri,lyw07/kolibri,MingDai/kolibri,rtibbles/kolibri,jonboiser/kolibri,indirectlylit/kolibri,jamalex/kolibri,whitzhu/kolibri,MingDai/kolibri,66eli77/kolibri,MingDai/kolibri,indirectlylit/kolibri,MingDai/kolibri,jamalex/kolibri,DXCanas/kolibri,jtamiace/kolibri,jamalex/kolibri,MCGallaspy/kolibri,christianmemije/kolibri,ralphiee22/kolibri,aronasorman/kolibri,aronasorman/kolibri,whitzhu/kolibri,jayoshih/kolibri,ralphiee22/kolibri,indirectlylit/kolibri,christianmemije/kolibri,learningequality/kolibri,benjaoming/kolibri,mrpau/kolibri,christianmemije/kolibri,ralphiee22/kolibri,jayoshih/kolibri | from __future__ import absolute_import, print_function, unicode_literals
+
+ import json
+ import tempfile
+
from django.test import TestCase
from ..hooks import WebpackBundleHook
+ TEST_STATS_FILE = None
+
+
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
- entry_file = "assets/src/kolibri_core_app.js"
+ src_file = "assets/src/kolibri_core_app.js"
+
+ @property
+ def stats_file(self):
+ return TEST_STATS_FILE.name
class KolibriTagNavigationTestCase(TestCase):
+ def setUp(self):
+ global TEST_STATS_FILE
+ TestCase.setUp(self)
+ TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False)
+ self.test_hook = TestHook()
+ json.dump(
+ {
+ "status": "done",
+ "chunks": {
+ "non_default_frontend": [
+ {
+ "name": "non_default_frontend-2c4fb3d6a29238b06f84.js",
+ "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js",
+ "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js"
+ }
+ ]
+ },
+ "publicPath": "default_frontend/"
+ },
+ TEST_STATS_FILE
+ )
+ TEST_STATS_FILE.close()
+
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
- TestHook().render_to_html()
+ self.test_hook.render_to_html()
)
| Create a faked hook that does not try to read a live stats file | ## Code Before:
from __future__ import absolute_import, print_function, unicode_literals
from django.test import TestCase
from ..hooks import WebpackBundleHook
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
entry_file = "assets/src/kolibri_core_app.js"
class KolibriTagNavigationTestCase(TestCase):
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
TestHook().render_to_html()
)
## Instruction:
Create a faked hook that does not try to read a live stats file
## Code After:
from __future__ import absolute_import, print_function, unicode_literals
import json
import tempfile
from django.test import TestCase
from ..hooks import WebpackBundleHook
TEST_STATS_FILE = None
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
src_file = "assets/src/kolibri_core_app.js"
@property
def stats_file(self):
return TEST_STATS_FILE.name
class KolibriTagNavigationTestCase(TestCase):
def setUp(self):
global TEST_STATS_FILE
TestCase.setUp(self)
TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False)
self.test_hook = TestHook()
json.dump(
{
"status": "done",
"chunks": {
"non_default_frontend": [
{
"name": "non_default_frontend-2c4fb3d6a29238b06f84.js",
"publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js",
"path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js"
}
]
},
"publicPath": "default_frontend/"
},
TEST_STATS_FILE
)
TEST_STATS_FILE.close()
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
self.test_hook.render_to_html()
)
| from __future__ import absolute_import, print_function, unicode_literals
+
+ import json
+ import tempfile
+
from django.test import TestCase
from ..hooks import WebpackBundleHook
+ TEST_STATS_FILE = None
+
+
class TestHook(WebpackBundleHook):
unique_slug = "non_default_frontend"
- entry_file = "assets/src/kolibri_core_app.js"
? ^^^ ^
+ src_file = "assets/src/kolibri_core_app.js"
? ^ ^
+
+ @property
+ def stats_file(self):
+ return TEST_STATS_FILE.name
class KolibriTagNavigationTestCase(TestCase):
+ def setUp(self):
+ global TEST_STATS_FILE
+ TestCase.setUp(self)
+ TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False)
+ self.test_hook = TestHook()
+ json.dump(
+ {
+ "status": "done",
+ "chunks": {
+ "non_default_frontend": [
+ {
+ "name": "non_default_frontend-2c4fb3d6a29238b06f84.js",
+ "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js",
+ "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js"
+ }
+ ]
+ },
+ "publicPath": "default_frontend/"
+ },
+ TEST_STATS_FILE
+ )
+ TEST_STATS_FILE.close()
+
def test_frontend_tag(self):
self.assertIn(
"non_default_frontend",
- TestHook().render_to_html()
? ^ ^ --
+ self.test_hook.render_to_html()
? ^^^^^^ ^^
) |
91f107ef2ebdaf7ff210b9f36e2c810441f389e7 | services/rdio.py | services/rdio.py | from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| Allow Rdio to use default signature handling | Allow Rdio to use default signature handling
| Python | bsd-3-clause | foauth/oauth-proxy,foauth/foauth.org,foauth/foauth.org,foauth/foauth.org | from werkzeug.urls import url_decode
- from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
- signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| Allow Rdio to use default signature handling | ## Code Before:
from werkzeug.urls import url_decode
from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
## Instruction:
Allow Rdio to use default signature handling
## Code After:
from werkzeug.urls import url_decode
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key'])
| from werkzeug.urls import url_decode
- from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY
import foauth.providers
class Rdio(foauth.providers.OAuth1):
# General info about the provider
provider_url = 'http://www.rdio.com/'
docs_url = 'http://developer.rdio.com/docs/REST/'
category = 'Music'
# URLs to interact with the API
request_token_url = 'http://api.rdio.com/oauth/request_token'
authorize_url = None # Provided when the request token is granted
access_token_url = 'http://api.rdio.com/oauth/access_token'
api_domain = 'api.rdio.com'
available_permissions = [
(None, 'access and manage your music'),
]
https = False
- signature_type = SIGNATURE_TYPE_BODY
def parse_token(self, content):
# Override standard token request to also get the authorization URL
data = url_decode(content)
if 'login_url' in data:
self.authorize_url = data['login_url']
return super(Rdio, self).parse_token(content)
def get_user_id(self, key):
r = self.api(key, self.api_domain, u'/1/', method='POST', data={
'method': 'currentUser',
})
return unicode(r.json[u'result'][u'key']) |
aeaf802100cd6869178dd9f412d35e452916a63d | common/commands/view_manipulation.py | common/commands/view_manipulation.py | from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
self.view.run_command("_enter_insert_mode")
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True)
| from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
# NeoVintageous renamed the command starting with v1.22.0.
# We call both commands for backwards compatibility.
self.view.run_command("_enter_insert_mode")
self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True)
| Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0 | Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0
Fixes #1395
In NeoVintageous/NeoVintageous#749, pushed as 1.22.0 (Oct 2020), the
relevant commands were renamed.
We follow the new names, but for now also call the old ones.
| Python | mit | divmain/GitSavvy,divmain/GitSavvy,divmain/GitSavvy | from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
+ # NeoVintageous renamed the command starting with v1.22.0.
+ # We call both commands for backwards compatibility.
self.view.run_command("_enter_insert_mode")
+ self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True)
| Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0 | ## Code Before:
from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
self.view.run_command("_enter_insert_mode")
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True)
## Instruction:
Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0
## Code After:
from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
# NeoVintageous renamed the command starting with v1.22.0.
# We call both commands for backwards compatibility.
self.view.run_command("_enter_insert_mode")
self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True)
| from sublime_plugin import TextCommand
from ...core.git_command import GitCommand
__all__ = (
"gs_handle_vintageous",
"gs_handle_arrow_keys"
)
class gs_handle_vintageous(TextCommand, GitCommand):
"""
Set the vintageous_friendly view setting if needed.
Enter insert mode if vintageous_enter_insert_mode option is enabled.
"""
def run(self, edit):
if self.savvy_settings.get("vintageous_friendly"):
self.view.settings().set("git_savvy.vintageous_friendly", True)
if self.savvy_settings.get("vintageous_enter_insert_mode"):
self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False)
+ # NeoVintageous renamed the command starting with v1.22.0.
+ # We call both commands for backwards compatibility.
self.view.run_command("_enter_insert_mode")
+ self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0
class gs_handle_arrow_keys(TextCommand, GitCommand):
"""
Set the arrow_keys_navigation view setting if needed.
It allows navigation by using arrow keys.
"""
def run(self, edit):
if self.savvy_settings.get("arrow_keys_navigation"):
self.view.settings().set("git_savvy.arrow_keys_navigation", True) |
4d414fe592bfd7f085f9aaea0b6992d28ad193ce | tcconfig/_common.py | tcconfig/_common.py |
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
"network interface not found: " + device)
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network
|
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
"network interface not found: {}".format(device))
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network
| Change to use format method | Change to use format method
| Python | mit | thombashi/tcconfig,thombashi/tcconfig |
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
- "network interface not found: " + device)
+ "network interface not found: {}".format(device))
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network
| Change to use format method | ## Code Before:
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
"network interface not found: " + device)
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network
## Instruction:
Change to use format method
## Code After:
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
"network interface not found: {}".format(device))
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network
|
from __future__ import absolute_import
import dataproperty
import six
from ._error import NetworkInterfaceNotFoundError
ANYWHERE_NETWORK = "0.0.0.0/0"
def verify_network_interface(device):
try:
import netifaces
except ImportError:
return
if device not in netifaces.interfaces():
raise NetworkInterfaceNotFoundError(
- "network interface not found: " + device)
? ^^^
+ "network interface not found: {}".format(device))
? ++ ^^^^^^^^ +
def sanitize_network(network):
"""
:return: Network string
:rtype: str
:raises ValueError: if the network string is invalid.
"""
import ipaddress
if dataproperty.is_empty_string(network):
return ""
if network.lower() == "anywhere":
return ANYWHERE_NETWORK
try:
ipaddress.IPv4Address(six.u(network))
return network + "/32"
except ipaddress.AddressValueError:
pass
ipaddress.IPv4Network(six.u(network)) # validate network str
return network |
5a03cd340e5dc8a796c7d430128f0e22be17333e | qiime/sdk/__init__.py | qiime/sdk/__init__.py |
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
|
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
# Various URLs
CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131'
HELP_URL = 'http://2.qiime.org'
CONDA_CHANNEL = 'https://anaconda.org/qiime2'
| Add helper URLs to qiime.sdk | ENH: Add helper URLs to qiime.sdk
Adds citation url, help page, and conda channel URLs to qiime.sdk
| Python | bsd-3-clause | biocore/qiime2,thermokarst/qiime2,ebolyen/qiime2,jakereps/qiime2,qiime2/qiime2,qiime2/qiime2,nervous-laughter/qiime2,biocore/qiime2,thermokarst/qiime2,jairideout/qiime2,jakereps/qiime2 |
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
+ # Various URLs
+ CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131'
+ HELP_URL = 'http://2.qiime.org'
+ CONDA_CHANNEL = 'https://anaconda.org/qiime2'
+ | Add helper URLs to qiime.sdk | ## Code Before:
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
## Instruction:
Add helper URLs to qiime.sdk
## Code After:
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
# Various URLs
CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131'
HELP_URL = 'http://2.qiime.org'
CONDA_CHANNEL = 'https://anaconda.org/qiime2'
|
from .method import Method
from .plugin_manager import PluginManager
from .provenance import Provenance
from .visualizer import Visualizer
from .result import Result, Artifact, Visualization
from ..core.util import parse_type
__all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer',
'PluginManager', 'Provenance', 'parse_type']
+
+ # Various URLs
+ CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131'
+ HELP_URL = 'http://2.qiime.org'
+ CONDA_CHANNEL = 'https://anaconda.org/qiime2' |
29fef644079a03fe0cfeb792dd47af7749382dba | unnaturalcode/http/__main__.py | unnaturalcode/http/__main__.py | from unnaturalcode.http import unnaturalhttp
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0')
|
try:
from unnaturalcode.http import unnaturalhttp
except ImportError:
import sys, os
# Oiugh.
sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
from unnaturalcode.http import unnaturalhttp
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0')
| Fix to allow invocation by `python unnaturalcode/http` | Fix to allow invocation by `python unnaturalcode/http`
| Python | agpl-3.0 | orezpraw/unnaturalcode,orezpraw/unnaturalcode,orezpraw/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,orezpraw/estimate-charm,naturalness/unnaturalcode,orezpraw/unnaturalcode | +
+ try:
- from unnaturalcode.http import unnaturalhttp
+ from unnaturalcode.http import unnaturalhttp
+ except ImportError:
+ import sys, os
+ # Oiugh.
+ sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
+ from unnaturalcode.http import unnaturalhttp
+
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0')
| Fix to allow invocation by `python unnaturalcode/http` | ## Code Before:
from unnaturalcode.http import unnaturalhttp
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0')
## Instruction:
Fix to allow invocation by `python unnaturalcode/http`
## Code After:
try:
from unnaturalcode.http import unnaturalhttp
except ImportError:
import sys, os
# Oiugh.
sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
from unnaturalcode.http import unnaturalhttp
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0')
| +
+ try:
- from unnaturalcode.http import unnaturalhttp
+ from unnaturalcode.http import unnaturalhttp
? ++++
+ except ImportError:
+ import sys, os
+ # Oiugh.
+ sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))
+ from unnaturalcode.http import unnaturalhttp
+
from flask import Flask
app = Flask(__name__)
app.register_blueprint(unnaturalhttp)
app.run(host='0.0.0.0') |
edc773bfd5d25a42fa2759631500fc4861557e57 | fireplace/cards/tgt/priest.py | fireplace/cards/tgt/priest.py | from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5)
| from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
# Shadowfiend
class AT_014:
events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e"))
# Wyrmrest Agent
class AT_116:
play = HOLDING_DRAGON & Buff(SELF, "AT_116e")
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5)
| Implement more TGT Priest cards | Implement more TGT Priest cards
| Python | agpl-3.0 | oftc-ftw/fireplace,beheh/fireplace,smallnamespace/fireplace,Meerkov/fireplace,Ragowit/fireplace,smallnamespace/fireplace,oftc-ftw/fireplace,amw2104/fireplace,liujimj/fireplace,NightKev/fireplace,jleclanche/fireplace,Ragowit/fireplace,amw2104/fireplace,liujimj/fireplace,Meerkov/fireplace | from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
+
+
+ # Shadowfiend
+ class AT_014:
+ events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e"))
+
+
+ # Wyrmrest Agent
+ class AT_116:
+ play = HOLDING_DRAGON & Buff(SELF, "AT_116e")
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5)
| Implement more TGT Priest cards | ## Code Before:
from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5)
## Instruction:
Implement more TGT Priest cards
## Code After:
from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
# Shadowfiend
class AT_014:
events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e"))
# Wyrmrest Agent
class AT_116:
play = HOLDING_DRAGON & Buff(SELF, "AT_116e")
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5)
| from ..utils import *
##
# Minions
# Holy Champion
class AT_011:
events = Heal().on(Buff(SELF, "AT_011e"))
# Spawn of Shadows
class AT_012:
inspire = Hit(ALL_HEROES, 4)
+
+
+ # Shadowfiend
+ class AT_014:
+ events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e"))
+
+
+ # Wyrmrest Agent
+ class AT_116:
+ play = HOLDING_DRAGON & Buff(SELF, "AT_116e")
##
# Spells
# Power Word: Glory
class AT_013:
play = Buff(TARGET, "AT_013e")
class AT_013e:
events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4))
# Convert
class AT_015:
play = Give(CONTROLLER, Copy(TARGET))
# Confuse
class AT_016:
play = Buff(ALL_MINIONS, "AT_016e")
# Flash Heal
class AT_055:
play = Heal(TARGET, 5) |
0b7f99bcb4e42c50263a7d8a42513876b02b445a | scikits/talkbox/tools/__init__.py | scikits/talkbox/tools/__init__.py | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
| __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| Put segment_axis in the main scikits.talkbox namespace. | Put segment_axis in the main scikits.talkbox namespace.
| Python | mit | cournape/talkbox,cournape/talkbox | __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
+ from segmentaxis import segment_axis
+ __all__ += ['segment_axis']
+ | Put segment_axis in the main scikits.talkbox namespace. | ## Code Before:
__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
## Instruction:
Put segment_axis in the main scikits.talkbox namespace.
## Code After:
__all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
from segmentaxis import segment_axis
__all__ += ['segment_axis']
| __all__ = []
import correlations
from correlations import *
__all__ += correlations.__all__
import cffilter
from cffilter import cslfilter as slfilter
__all__ += ['slfilter']
+
+ from segmentaxis import segment_axis
+ __all__ += ['segment_axis'] |
df8f5e0a6be5f3de31d61810b1624175b2d105ec | auth0/v2/device_credentials.py | auth0/v2/device_credentials.py | from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
| from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
def create(self, body):
return self.client.post(data=body)
def delete(self, id):
return self.client.delete(id=id)
| Implement create and delete methods for DeviceCredentials | Implement create and delete methods for DeviceCredentials
| Python | mit | auth0/auth0-python,auth0/auth0-python | from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
+ def create(self, body):
+ return self.client.post(data=body)
+
+ def delete(self, id):
+ return self.client.delete(id=id)
+ | Implement create and delete methods for DeviceCredentials | ## Code Before:
from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
## Instruction:
Implement create and delete methods for DeviceCredentials
## Code After:
from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
def create(self, body):
return self.client.post(data=body)
def delete(self, id):
return self.client.delete(id=id)
| from .rest import RestClient
class DeviceCredentials(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/device-credentials' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, user_id=None, client_id=None, type=None,
fields=[], include_fields=True):
params = {
'fields': ','.join(fields) or None,
'include_fields': str(include_fields).lower(),
'user_id': user_id,
'client_id': client_id,
'type': type,
}
return self.client.get(params=params)
+
+ def create(self, body):
+ return self.client.post(data=body)
+
+ def delete(self, id):
+ return self.client.delete(id=id) |
278b17859e4ad7464098a715777fcb755acf258c | doTranscode.py | doTranscode.py | import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile()
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
| import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile(suffix=".wav")
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
| Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file | Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file | Python | isc | jeffayle/Transcode | import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
- f = tempfile.NamedTemporaryFile()
+ f = tempfile.NamedTemporaryFile(suffix=".wav")
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
| Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file | ## Code Before:
import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile()
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
## Instruction:
Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file
## Code After:
import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
f = tempfile.NamedTemporaryFile(suffix=".wav")
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ
| import encoders
import decoders
import config
import tempfile
import os
def transcode(inF, outF, options, type=None):
"Transcodes a file"
if type == None:
type = os.path.splitext(outF)[1][1:].lower()
#Get the file's metadata
meta = decoders.getMetadata(inF)
#Decode the file
- f = tempfile.NamedTemporaryFile()
+ f = tempfile.NamedTemporaryFile(suffix=".wav")
? +++++++++++++
inF_real = decoders.decode(inF, f.name)
if not inF_real:
return False
#Encode it
succ = encoders.encode(inF_real, outF, type, options, meta)
#Clean up
f.close()
return succ |
0a13a9a8a779102dbcb2beead7d8aa9143f4c79b | tests/pytests/unit/client/ssh/test_shell.py | tests/pytests/unit/client/ssh/test_shell.py | import os
import subprocess
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
class TestSSHShell:
def test_ssh_shell_key_gen(self, keys):
"""
Test ssh key_gen
"""
shell.gen_key(keys["priv_key"])
for fp in keys.keys():
assert os.path.exists(keys[fp])
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
| import subprocess
import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
def test_ssh_shell_key_gen(keys):
"""
Test ssh key_gen
"""
shell.gen_key(str(keys.priv_key))
assert keys.priv_key.exists()
assert keys.pub_key.exists()
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
| Use commit suggestion to use types | Use commit suggestion to use types
Co-authored-by: Pedro Algarvio <4410d99cefe57ec2c2cdbd3f1d5cf862bb4fb6f8@algarvio.me>
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | - import os
import subprocess
+ import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
- yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
+ return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
- class TestSSHShell:
- def test_ssh_shell_key_gen(self, keys):
+ def test_ssh_shell_key_gen(keys):
- """
+ """
- Test ssh key_gen
+ Test ssh key_gen
- """
+ """
- shell.gen_key(keys["priv_key"])
+ shell.gen_key(str(keys.priv_key))
- for fp in keys.keys():
- assert os.path.exists(keys[fp])
+ assert keys.priv_key.exists()
+ assert keys.pub_key.exists()
+ # verify there is not a passphrase set on key
+ ret = subprocess.check_output(
+ ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
+ )
+ assert ret.decode().startswith("ssh-rsa")
- # verify there is not a passphrase set on key
- ret = subprocess.check_output(
- ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
- )
- assert ret.decode().startswith("ssh-rsa")
- | Use commit suggestion to use types | ## Code Before:
import os
import subprocess
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
class TestSSHShell:
def test_ssh_shell_key_gen(self, keys):
"""
Test ssh key_gen
"""
shell.gen_key(keys["priv_key"])
for fp in keys.keys():
assert os.path.exists(keys[fp])
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
## Instruction:
Use commit suggestion to use types
## Code After:
import subprocess
import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
def test_ssh_shell_key_gen(keys):
"""
Test ssh key_gen
"""
shell.gen_key(str(keys.priv_key))
assert keys.priv_key.exists()
assert keys.pub_key.exists()
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
| - import os
import subprocess
+ import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
- yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
+ return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
- class TestSSHShell:
- def test_ssh_shell_key_gen(self, keys):
? ---- ------
+ def test_ssh_shell_key_gen(keys):
- """
? ----
+ """
- Test ssh key_gen
? ----
+ Test ssh key_gen
- """
? ----
+ """
- shell.gen_key(keys["priv_key"])
? ---- ^^ --
+ shell.gen_key(str(keys.priv_key))
? ++++ ^ +
+ assert keys.priv_key.exists()
+ assert keys.pub_key.exists()
- for fp in keys.keys():
- assert os.path.exists(keys[fp])
-
- # verify there is not a passphrase set on key
? ----
+ # verify there is not a passphrase set on key
- ret = subprocess.check_output(
? ----
+ ret = subprocess.check_output(
- ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
? ---- ^^ ^^
+ ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
? ++++ ^ ^
- )
+ )
- assert ret.decode().startswith("ssh-rsa")
? ----
+ assert ret.decode().startswith("ssh-rsa") |
c668aaa0f22f5a61094c2028291b65c781733a54 | mojapi/api.py | mojapi/api.py | import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
| import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
def get_blocked_server_hashes():
response = requests.get('https://sessionserver.mojang.com/blockedservers')
response.raise_for_status()
sha1_hashes = response.content.split(b'\n')
return sha1_hashes
| Add get blocked server hashes call | Add get blocked server hashes call
| Python | mit | zugmc/mojapi | import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
+
+ def get_blocked_server_hashes():
+ response = requests.get('https://sessionserver.mojang.com/blockedservers')
+ response.raise_for_status()
+ sha1_hashes = response.content.split(b'\n')
+ return sha1_hashes
+ | Add get blocked server hashes call | ## Code Before:
import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
## Instruction:
Add get blocked server hashes call
## Code After:
import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
def get_blocked_server_hashes():
response = requests.get('https://sessionserver.mojang.com/blockedservers')
response.raise_for_status()
sha1_hashes = response.content.split(b'\n')
return sha1_hashes
| import json
import requests
import time
def get_statuses():
return requests.get('https://status.mojang.com/check/').json()
def get_uuid(username, unix_timestamp=None):
if unix_timestamp is None:
unix_timestamp = int(time.time())
return requests.get(
'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp)
).json()
def get_usernames(uuid):
return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json()
def get_profiles(*usernames):
return requests.post(
url='https://api.mojang.com/profiles/minecraft',
headers={
b'Content-Type': b'application/json'
},
data=json.dumps(list(usernames))
).json()
+
+
+ def get_blocked_server_hashes():
+ response = requests.get('https://sessionserver.mojang.com/blockedservers')
+ response.raise_for_status()
+ sha1_hashes = response.content.split(b'\n')
+ return sha1_hashes |
5dec1db567ef7c2b6ea1cca3ddd02612cb9f7d8a | Lib/encodings/bz2_codec.py | Lib/encodings/bz2_codec.py | import codecs
import bz2
def encode(input, errors='strict'):
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
def decode(input, errors='strict'):
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
### encodings module API
def getregentry():
return (encode, decode, codecs.StreamReader, codecs.StreamWriter)
| import codecs
import bz2 # this codec needs the optional bz2 module !
### Codec APIs
def bz2_encode(input,errors='strict'):
""" Encodes the object input and returns a tuple (output
object, length consumed).
errors defines the error handling to apply. It defaults to
'strict' handling which is the only currently supported
error handling for this codec.
"""
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
def bz2_decode(input,errors='strict'):
""" Decodes the object input and returns a tuple (output
object, length consumed).
input must be an object which provides the bf_getreadbuf
buffer slot. Python strings, buffer objects and memory
mapped files are examples of objects providing this slot.
errors defines the error handling to apply. It defaults to
'strict' handling which is the only currently supported
error handling for this codec.
"""
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
class Codec(codecs.Codec):
def encode(self, input, errors='strict'):
return bz2_encode(input, errors)
def decode(self, input, errors='strict'):
return bz2_decode(input, errors)
class StreamWriter(Codec,codecs.StreamWriter):
pass
class StreamReader(Codec,codecs.StreamReader):
pass
### encodings module API
def getregentry():
return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
| Revert previous change. MAL preferred the old version. | Revert previous change. MAL preferred the old version.
| Python | mit | sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator | import codecs
- import bz2
+ import bz2 # this codec needs the optional bz2 module !
+ ### Codec APIs
+
- def encode(input, errors='strict'):
+ def bz2_encode(input,errors='strict'):
+
+ """ Encodes the object input and returns a tuple (output
+ object, length consumed).
+
+ errors defines the error handling to apply. It defaults to
+ 'strict' handling which is the only currently supported
+ error handling for this codec.
+
+ """
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
- def decode(input, errors='strict'):
+ def bz2_decode(input,errors='strict'):
+
+ """ Decodes the object input and returns a tuple (output
+ object, length consumed).
+
+ input must be an object which provides the bf_getreadbuf
+ buffer slot. Python strings, buffer objects and memory
+ mapped files are examples of objects providing this slot.
+
+ errors defines the error handling to apply. It defaults to
+ 'strict' handling which is the only currently supported
+ error handling for this codec.
+
+ """
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
+
+ class Codec(codecs.Codec):
+
+ def encode(self, input, errors='strict'):
+ return bz2_encode(input, errors)
+ def decode(self, input, errors='strict'):
+ return bz2_decode(input, errors)
+
+ class StreamWriter(Codec,codecs.StreamWriter):
+ pass
+
+ class StreamReader(Codec,codecs.StreamReader):
+ pass
### encodings module API
def getregentry():
- return (encode, decode, codecs.StreamReader, codecs.StreamWriter)
+ return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
| Revert previous change. MAL preferred the old version. | ## Code Before:
import codecs
import bz2
def encode(input, errors='strict'):
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
def decode(input, errors='strict'):
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
### encodings module API
def getregentry():
return (encode, decode, codecs.StreamReader, codecs.StreamWriter)
## Instruction:
Revert previous change. MAL preferred the old version.
## Code After:
import codecs
import bz2 # this codec needs the optional bz2 module !
### Codec APIs
def bz2_encode(input,errors='strict'):
""" Encodes the object input and returns a tuple (output
object, length consumed).
errors defines the error handling to apply. It defaults to
'strict' handling which is the only currently supported
error handling for this codec.
"""
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
def bz2_decode(input,errors='strict'):
""" Decodes the object input and returns a tuple (output
object, length consumed).
input must be an object which provides the bf_getreadbuf
buffer slot. Python strings, buffer objects and memory
mapped files are examples of objects providing this slot.
errors defines the error handling to apply. It defaults to
'strict' handling which is the only currently supported
error handling for this codec.
"""
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
class Codec(codecs.Codec):
def encode(self, input, errors='strict'):
return bz2_encode(input, errors)
def decode(self, input, errors='strict'):
return bz2_decode(input, errors)
class StreamWriter(Codec,codecs.StreamWriter):
pass
class StreamReader(Codec,codecs.StreamReader):
pass
### encodings module API
def getregentry():
return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
| import codecs
- import bz2
+ import bz2 # this codec needs the optional bz2 module !
+ ### Codec APIs
+
- def encode(input, errors='strict'):
? -
+ def bz2_encode(input,errors='strict'):
? ++++
+
+ """ Encodes the object input and returns a tuple (output
+ object, length consumed).
+
+ errors defines the error handling to apply. It defaults to
+ 'strict' handling which is the only currently supported
+ error handling for this codec.
+
+ """
assert errors == 'strict'
output = bz2.compress(input)
return (output, len(input))
- def decode(input, errors='strict'):
? -
+ def bz2_decode(input,errors='strict'):
? ++++
+
+ """ Decodes the object input and returns a tuple (output
+ object, length consumed).
+
+ input must be an object which provides the bf_getreadbuf
+ buffer slot. Python strings, buffer objects and memory
+ mapped files are examples of objects providing this slot.
+
+ errors defines the error handling to apply. It defaults to
+ 'strict' handling which is the only currently supported
+ error handling for this codec.
+
+ """
assert errors == 'strict'
output = bz2.decompress(input)
return (output, len(input))
+
+ class Codec(codecs.Codec):
+
+ def encode(self, input, errors='strict'):
+ return bz2_encode(input, errors)
+ def decode(self, input, errors='strict'):
+ return bz2_decode(input, errors)
+
+ class StreamWriter(Codec,codecs.StreamWriter):
+ pass
+
+ class StreamReader(Codec,codecs.StreamReader):
+ pass
### encodings module API
def getregentry():
- return (encode, decode, codecs.StreamReader, codecs.StreamWriter)
? ^ -------- --------
+ return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
? ++++ ^^^^
|
f22476a36f2096628dc336f9adf0caa9a827dc11 | jfr_playoff/db.py | jfr_playoff/db.py | import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
| import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
raise IOError(e.errno, str(e), db_name)
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
| Fix for rethrowing mysql.connector.Error as IOError | Fix for rethrowing mysql.connector.Error as IOError
| Python | bsd-2-clause | emkael/jfrteamy-playoff,emkael/jfrteamy-playoff | import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
+ raise IOError(e.errno, str(e), db_name)
- raise IOError(
- message=str(e), filename=db_name,
- errno=e.errno, strerror=str(e))
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
| Fix for rethrowing mysql.connector.Error as IOError | ## Code Before:
import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
## Instruction:
Fix for rethrowing mysql.connector.Error as IOError
## Code After:
import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
raise IOError(e.errno, str(e), db_name)
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e))
| import sys
class PlayoffDB(object):
db_cursor = None
DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured'
def __init__(self, settings):
reload(sys)
sys.setdefaultencoding("latin1")
import mysql.connector
self.database = mysql.connector.connect(
user=settings['user'],
password=settings['pass'],
host=settings['host'],
port=settings['port'])
self.db_cursor = self.database.cursor(buffered=True)
def get_cursor(self):
return self.db_cursor
def __execute_query(self, db_name, sql, params):
self.db_cursor.execute(sql.replace('#db#', db_name), params)
def fetch(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
row = self.db_cursor.fetchone()
return row
except mysql.connector.Error as e:
+ raise IOError(e.errno, str(e), db_name)
- raise IOError(
- message=str(e), filename=db_name,
- errno=e.errno, strerror=str(e))
def fetch_all(self, db_name, sql, params):
import mysql.connector
try:
self.__execute_query(db_name, sql, params)
results = self.db_cursor.fetchall()
return results
except mysql.connector.Error as e:
raise IOError(
message=str(e), filename=db_name,
errno=e.errno, strerror=str(e)) |
7917716ebd11770c5d4d0634b39e32e4f577ab71 | tests/test_urls.py | tests/test_urls.py | from unittest import TestCase
class TestURLs(TestCase):
pass
| from unittest import TestCase
from django.contrib.auth import views
from django.core.urlresolvers import resolve, reverse
class URLsMixin(object):
"""
A TestCase Mixin with a check_url helper method for testing urls.
Pirated with slight modifications from incuna_test_utils
https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py
"""
def check_url(self, view_method, expected_url, url_name,
url_args=None, url_kwargs=None):
"""
Assert a view's url is correctly configured
Check the url_name reverses to give a correctly formated expected_url.
Check the expected_url resolves to the correct view.
"""
reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)
self.assertEqual(reversed_url, expected_url)
# Look for a method rather than a class here
# (just because of what we're testing)
resolved_view_method = resolve(expected_url).func
self.assertEqual(resolved_view_method, view_method)
class TestURLs(URLsMixin, TestCase):
def test_login(self):
self.check_url(
views.login,
'/login/',
'login',
)
def test_logout(self):
self.check_url(
views.logout,
'/logout/',
'logout',
)
def test_password_change(self):
self.check_url(
views.password_change,
'/password/change/',
'password_change',
)
def test_password_change_done(self):
self.check_url(
views.password_change_done,
'/password/change/done/',
'password_change_done',
)
def test_password_reset(self):
self.check_url(
views.password_reset,
'/password/reset/',
'password_reset',
)
def test_password_reset_done(self):
self.check_url(
views.password_reset_done,
'/password/reset/done/',
'password_reset_done',
)
def test_password_reset_complete(self):
self.check_url(
views.password_reset_complete,
'/password/reset/complete/',
'password_reset_complete',
)
| Add lots of URL tests. | Add lots of URL tests.
* The URLsMixin from incuna_test_utils/testcases/urls.py
isn't quite doing what we want here, so rip it off and
make a small modification (resolve(...).func.cls ->
resolve(...).func).
* Add lots of tests for the django.contrib.auth views
that we're using (the others are more complex).
| Python | bsd-2-clause | incuna/incuna-auth,ghickman/incuna-auth,ghickman/incuna-auth,incuna/incuna-auth | from unittest import TestCase
+ from django.contrib.auth import views
+ from django.core.urlresolvers import resolve, reverse
- class TestURLs(TestCase):
- pass
+ class URLsMixin(object):
+ """
+ A TestCase Mixin with a check_url helper method for testing urls.
+ Pirated with slight modifications from incuna_test_utils
+ https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py
+ """
+
+ def check_url(self, view_method, expected_url, url_name,
+ url_args=None, url_kwargs=None):
+ """
+ Assert a view's url is correctly configured
+
+ Check the url_name reverses to give a correctly formated expected_url.
+ Check the expected_url resolves to the correct view.
+ """
+
+ reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)
+ self.assertEqual(reversed_url, expected_url)
+
+ # Look for a method rather than a class here
+ # (just because of what we're testing)
+ resolved_view_method = resolve(expected_url).func
+ self.assertEqual(resolved_view_method, view_method)
+
+
+ class TestURLs(URLsMixin, TestCase):
+
+ def test_login(self):
+ self.check_url(
+ views.login,
+ '/login/',
+ 'login',
+ )
+
+ def test_logout(self):
+ self.check_url(
+ views.logout,
+ '/logout/',
+ 'logout',
+ )
+
+ def test_password_change(self):
+ self.check_url(
+ views.password_change,
+ '/password/change/',
+ 'password_change',
+ )
+
+ def test_password_change_done(self):
+ self.check_url(
+ views.password_change_done,
+ '/password/change/done/',
+ 'password_change_done',
+ )
+
+ def test_password_reset(self):
+ self.check_url(
+ views.password_reset,
+ '/password/reset/',
+ 'password_reset',
+ )
+
+ def test_password_reset_done(self):
+ self.check_url(
+ views.password_reset_done,
+ '/password/reset/done/',
+ 'password_reset_done',
+ )
+
+ def test_password_reset_complete(self):
+ self.check_url(
+ views.password_reset_complete,
+ '/password/reset/complete/',
+ 'password_reset_complete',
+ )
+ | Add lots of URL tests. | ## Code Before:
from unittest import TestCase
class TestURLs(TestCase):
pass
## Instruction:
Add lots of URL tests.
## Code After:
from unittest import TestCase
from django.contrib.auth import views
from django.core.urlresolvers import resolve, reverse
class URLsMixin(object):
"""
A TestCase Mixin with a check_url helper method for testing urls.
Pirated with slight modifications from incuna_test_utils
https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py
"""
def check_url(self, view_method, expected_url, url_name,
url_args=None, url_kwargs=None):
"""
Assert a view's url is correctly configured
Check the url_name reverses to give a correctly formated expected_url.
Check the expected_url resolves to the correct view.
"""
reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)
self.assertEqual(reversed_url, expected_url)
# Look for a method rather than a class here
# (just because of what we're testing)
resolved_view_method = resolve(expected_url).func
self.assertEqual(resolved_view_method, view_method)
class TestURLs(URLsMixin, TestCase):
def test_login(self):
self.check_url(
views.login,
'/login/',
'login',
)
def test_logout(self):
self.check_url(
views.logout,
'/logout/',
'logout',
)
def test_password_change(self):
self.check_url(
views.password_change,
'/password/change/',
'password_change',
)
def test_password_change_done(self):
self.check_url(
views.password_change_done,
'/password/change/done/',
'password_change_done',
)
def test_password_reset(self):
self.check_url(
views.password_reset,
'/password/reset/',
'password_reset',
)
def test_password_reset_done(self):
self.check_url(
views.password_reset_done,
'/password/reset/done/',
'password_reset_done',
)
def test_password_reset_complete(self):
self.check_url(
views.password_reset_complete,
'/password/reset/complete/',
'password_reset_complete',
)
| from unittest import TestCase
+ from django.contrib.auth import views
+ from django.core.urlresolvers import resolve, reverse
+
+ class URLsMixin(object):
+ """
+ A TestCase Mixin with a check_url helper method for testing urls.
+ Pirated with slight modifications from incuna_test_utils
+ https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py
+ """
+
+ def check_url(self, view_method, expected_url, url_name,
+ url_args=None, url_kwargs=None):
+ """
+ Assert a view's url is correctly configured
+
+ Check the url_name reverses to give a correctly formated expected_url.
+ Check the expected_url resolves to the correct view.
+ """
+
+ reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)
+ self.assertEqual(reversed_url, expected_url)
+
+ # Look for a method rather than a class here
+ # (just because of what we're testing)
+ resolved_view_method = resolve(expected_url).func
+ self.assertEqual(resolved_view_method, view_method)
+
+
- class TestURLs(TestCase):
+ class TestURLs(URLsMixin, TestCase):
? +++++++++++
- pass
+
+ def test_login(self):
+ self.check_url(
+ views.login,
+ '/login/',
+ 'login',
+ )
+
+ def test_logout(self):
+ self.check_url(
+ views.logout,
+ '/logout/',
+ 'logout',
+ )
+
+ def test_password_change(self):
+ self.check_url(
+ views.password_change,
+ '/password/change/',
+ 'password_change',
+ )
+
+ def test_password_change_done(self):
+ self.check_url(
+ views.password_change_done,
+ '/password/change/done/',
+ 'password_change_done',
+ )
+
+ def test_password_reset(self):
+ self.check_url(
+ views.password_reset,
+ '/password/reset/',
+ 'password_reset',
+ )
+
+ def test_password_reset_done(self):
+ self.check_url(
+ views.password_reset_done,
+ '/password/reset/done/',
+ 'password_reset_done',
+ )
+
+ def test_password_reset_complete(self):
+ self.check_url(
+ views.password_reset_complete,
+ '/password/reset/complete/',
+ 'password_reset_complete',
+ ) |
6fecc53b63023e6d25722aa66038285be3b4d46b | arcutils/response.py | arcutils/response.py | from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
location = (
request.GET.get(redirect_field_name) or
request.POST.get(redirect_field_name) or
request.META.get('HTTP_REFERER')
)
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
return location
| from urllib.parse import urlparse, urlunparse
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name)
if location:
from_referrer = False
else:
location = request.META.get('HTTP_REFERER')
from_referrer = bool(location)
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
elif from_referrer:
info = urlparse(location)
if info.netloc == host:
# Clear scheme and host (AKA netloc) to get just the path of
# the referrer. Also, ensure the path is set for consistency.
new_info = ('', '', info.path or '/') + info[3:]
location = urlunparse(new_info)
return location
| Return just path when getting redirect location from REFERER | Return just path when getting redirect location from REFERER
In response.get_redirect_location(). There's no need to include the
scheme and host in a redirect back to the same site. Removing them makes
redirect URLs more concise.
| Python | mit | PSU-OIT-ARC/django-arcutils,wylee/django-arcutils,PSU-OIT-ARC/django-arcutils,wylee/django-arcutils | + from urllib.parse import urlparse, urlunparse
+
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
+ location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name)
+
- location = (
+ if location:
- request.GET.get(redirect_field_name) or
- request.POST.get(redirect_field_name) or
+ from_referrer = False
+ else:
- request.META.get('HTTP_REFERER')
+ location = request.META.get('HTTP_REFERER')
- )
+ from_referrer = bool(location)
+
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
+ elif from_referrer:
+ info = urlparse(location)
+ if info.netloc == host:
+ # Clear scheme and host (AKA netloc) to get just the path of
+ # the referrer. Also, ensure the path is set for consistency.
+ new_info = ('', '', info.path or '/') + info[3:]
+ location = urlunparse(new_info)
+
return location
| Return just path when getting redirect location from REFERER | ## Code Before:
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
location = (
request.GET.get(redirect_field_name) or
request.POST.get(redirect_field_name) or
request.META.get('HTTP_REFERER')
)
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
return location
## Instruction:
Return just path when getting redirect location from REFERER
## Code After:
from urllib.parse import urlparse, urlunparse
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name)
if location:
from_referrer = False
else:
location = request.META.get('HTTP_REFERER')
from_referrer = bool(location)
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
elif from_referrer:
info = urlparse(location)
if info.netloc == host:
# Clear scheme and host (AKA netloc) to get just the path of
# the referrer. Also, ensure the path is set for consistency.
new_info = ('', '', info.path or '/') + info[3:]
location = urlunparse(new_info)
return location
| + from urllib.parse import urlparse, urlunparse
+
from django.contrib.auth import REDIRECT_FIELD_NAME
from django.utils.http import is_safe_url
def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'):
"""Attempt to choose an optimal redirect location.
If a location is specified via a request parameter, that location
will be used.
If a location is specified via POST or PUT data, that location will
be used.
In either of the above two cases, the name of the parameter and data
field is specified by ``redirect_field_name``, which defaults to
"next".
Otherwise, the preferred option is to redirect back to the referring
page.
If there's no referrer, the default is used.
In any case, the redirect location must be safe (same host, safe
scheme). Otherwise, the ``default`` location will be used. If the
default location isn't safe, "/" will be used as a last resort.
"""
host = request.get_host()
+ location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name)
+
- location = (
? ^^^^
+ if location:
? +++ ^
- request.GET.get(redirect_field_name) or
- request.POST.get(redirect_field_name) or
+ from_referrer = False
+ else:
- request.META.get('HTTP_REFERER')
+ location = request.META.get('HTTP_REFERER')
? +++++++++++
- )
+ from_referrer = bool(location)
+
if not is_safe_url(location, host):
default = default or '/'
if not is_safe_url(default, host):
default = '/'
location = default
+ elif from_referrer:
+ info = urlparse(location)
+ if info.netloc == host:
+ # Clear scheme and host (AKA netloc) to get just the path of
+ # the referrer. Also, ensure the path is set for consistency.
+ new_info = ('', '', info.path or '/') + info[3:]
+ location = urlunparse(new_info)
+
return location |
456b72757cda81c8dd6634ae41b8a1008ff59087 | config-example.py | config-example.py |
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
]
|
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
# Use this picture as a watermark, for sponsorships etc. Should be a PNG image
# with transparency. It is overlaid directly with the camera image, so it
# should have the same dimensions as `camera_dimensions` above. Leave as an
# empty string to have no watermark.
watermark_path = ""
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
]
| Add watermark path to example config | Add watermark path to example config
| Python | mit | mgunyho/kiltiskahvi |
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
+
+ # Use this picture as a watermark, for sponsorships etc. Should be a PNG image
+ # with transparency. It is overlaid directly with the camera image, so it
+ # should have the same dimensions as `camera_dimensions` above. Leave as an
+ # empty string to have no watermark.
+ watermark_path = ""
+
+
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
]
| Add watermark path to example config | ## Code Before:
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
]
## Instruction:
Add watermark path to example config
## Code After:
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
# Use this picture as a watermark, for sponsorships etc. Should be a PNG image
# with transparency. It is overlaid directly with the camera image, so it
# should have the same dimensions as `camera_dimensions` above. Leave as an
# empty string to have no watermark.
watermark_path = ""
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
]
|
# put your bot token here as a string
bot_token = ""
# the tg username of the bot's admin.
admin_username = ""
# The size of the pictures the webcamera takes. As of 2022-03-06, the guild
# room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to
# be 720p
camera_dimensions = (1280, 720)
+
+ # Use this picture as a watermark, for sponsorships etc. Should be a PNG image
+ # with transparency. It is overlaid directly with the camera image, so it
+ # should have the same dimensions as `camera_dimensions` above. Leave as an
+ # empty string to have no watermark.
+ watermark_path = ""
+
+
# if a message contains any of these words, the bot responds
trigger_words = [
"kahvi",
"\u2615", # coffee emoji
"tsufe",
"kahavi",
#"sima", # wappu mode
] |
59ce3ca9c1572dcf71aa5de5cdb354def594a36c | downloads/urls.py | downloads/urls.py | from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
url(r'solutions/(?P<path>.*)$', partial(download_protected_file,
path_prefix='solutions/',
model_class=UserSolution),
name='download_solution'),
)
| from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
url(r'solutions/(?P<path>.*)$', download_protected_file,
dict(path_prefix='solutions/', model_class=UserSolution),
name='download_solution'),
)
| Remove unnecessary usage of functools.partial | downloads: Remove unnecessary usage of functools.partial
| Python | mit | matus-stehlik/roots,rtrembecky/roots,matus-stehlik/roots,tbabej/roots,rtrembecky/roots,tbabej/roots,tbabej/roots,rtrembecky/roots,matus-stehlik/roots | from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
- url(r'solutions/(?P<path>.*)$', partial(download_protected_file,
+ url(r'solutions/(?P<path>.*)$', download_protected_file,
+ dict(path_prefix='solutions/', model_class=UserSolution),
- path_prefix='solutions/',
- model_class=UserSolution),
name='download_solution'),
)
| Remove unnecessary usage of functools.partial | ## Code Before:
from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
url(r'solutions/(?P<path>.*)$', partial(download_protected_file,
path_prefix='solutions/',
model_class=UserSolution),
name='download_solution'),
)
## Instruction:
Remove unnecessary usage of functools.partial
## Code After:
from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
url(r'solutions/(?P<path>.*)$', download_protected_file,
dict(path_prefix='solutions/', model_class=UserSolution),
name='download_solution'),
)
| from django.conf.urls import patterns, url
from functools import partial
from problems.models import UserSolution
from .views import download_protected_file
urlpatterns = patterns('',
- url(r'solutions/(?P<path>.*)$', partial(download_protected_file,
? --------
+ url(r'solutions/(?P<path>.*)$', download_protected_file,
+ dict(path_prefix='solutions/', model_class=UserSolution),
- path_prefix='solutions/',
- model_class=UserSolution),
name='download_solution'),
) |
59afb96f2211983ee2a2786c60791074b13c3e7f | ni/__main__.py | ni/__main__.py | """Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
# With a background queue, one could add the update as a work item
# and return an HTTP 202 response.
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port())
| """Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
# With a work queue, one could make the updating of the
# contribution a work item and return an HTTP 202 response.
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port())
| Tweak comment about 202 response | Tweak comment about 202 response
| Python | apache-2.0 | python/the-knights-who-say-ni,python/the-knights-who-say-ni | """Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
- # With a background queue, one could add the update as a work item
+ # With a work queue, one could make the updating of the
- # and return an HTTP 202 response.
+ # contribution a work item and return an HTTP 202 response.
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port())
| Tweak comment about 202 response | ## Code Before:
"""Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
# With a background queue, one could add the update as a work item
# and return an HTTP 202 response.
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port())
## Instruction:
Tweak comment about 202 response
## Code After:
"""Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
# With a work queue, one could make the updating of the
# contribution a work item and return an HTTP 202 response.
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port())
| """Implement a server to check if a contribution is covered by a CLA(s)."""
from aiohttp import web
from . import abc
from . import ContribHost
from . import ServerHost
from . import CLAHost
class Handler:
"""Handle requests from the contribution host."""
def __init__(self, server: ServerHost, cla_records: CLAHost):
self.server = server
self.cla_records = cla_records
async def respond(request: web.Request) -> web.StreamResponse: # XXX untested
"""Handle a webhook trigger from the contribution host."""
try:
contribution = ContribHost.process(request)
usernames = await contribution.usernames() # XXX not implemented
cla_status = await self.cla_records.check(usernames) # XXX not implemented
- # With a background queue, one could add the update as a work item
+ # With a work queue, one could make the updating of the
- # and return an HTTP 202 response.
+ # contribution a work item and return an HTTP 202 response.
? +++++++++++++++++++++++++
return (await contribution.update(cla_status)) # XXX not implemented
except abc.ResponseExit as exc:
return exc.response
except Exception as exc:
self.server.log(exc)
return web.Response(
status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value)
if __name__ == '__main__':
server = ServerHost()
cla_records = CLAHost()
handler = Handler(server, cla_records)
app = web.Application()
app.router.add_route(*ContribHost.route, handler.respond)
web.run_app(app, port=server.port()) |
2814d7b8060d1f468bb6fb34d1460cdad1811031 | tools/android/emulator/reporting.py | tools/android/emulator/reporting.py | """An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
unused_runtime_ms, unused_success):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
| """An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
unused_runtime_ms, unused_success,
unused_total_runtime):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
| Update the reporter interface to even track the total runtime | Update the reporter interface to even track the total runtime
PiperOrigin-RevId: 160982468
| Python | apache-2.0 | android/android-test,android/android-test,android/android-test,android/android-test,android/android-test | """An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
- unused_runtime_ms, unused_success):
+ unused_runtime_ms, unused_success,
+ unused_total_runtime):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
| Update the reporter interface to even track the total runtime | ## Code Before:
"""An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
unused_runtime_ms, unused_success):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
## Instruction:
Update the reporter interface to even track the total runtime
## Code After:
"""An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
unused_runtime_ms, unused_success,
unused_total_runtime):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
| """An interface to report the status of emulator launches."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import json
import logging
import os
import uuid
class NoOpReporter(object):
"""Captures all device and failure data and throws it away."""
def __init__(self):
pass
def ReportDeviceProperties(self, unused_emu_type, unused_props):
pass
def ReportFailure(self, unused_component, unused_details):
pass
def ReportToolsUsage(self, unused_namespace, unused_tool_name,
- unused_runtime_ms, unused_success):
? ^^
+ unused_runtime_ms, unused_success,
? ^
+ unused_total_runtime):
pass
def Emit(self):
pass
def MakeReporter():
"""Creates a reporter instance."""
return NoOpReporter()
|
a85beb35d7296b0a8bd5a385b44fa13fb9f178ed | imgur-clean.py | imgur-clean.py |
import hashlib
import re
import os
import sys
IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)')
def get_hash(fn):
with open(fn, 'rb') as fh:
hashsum = hashlib.md5(fh.read()).digest()
return hashsum
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
sums = {}
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
new_fn = f'{match.group(1)}.{match.group(2)}'
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
hashsum = get_hash(new_fn)
files = sums.get(hashsum, [])
files.append(new_fn)
sums[hashsum] = files
for hashsum, files in sums.items():
if len(files) > 1:
files_quoted = [f"'{x}'" for x in files]
print(f"Found duplicates: {', '.join(files_quoted)}")
files.sort()
for fn in files[1:]:
os.remove(fn)
|
import re
import os
import sys
IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)')
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
ids = {}
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
new_fn = f'{match[1]}.{match[3]}'
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
id = match[2]
files = ids.get(id, [])
files.append(new_fn)
ids[id] = files
for _, files in ids.items():
if len(files) > 1:
files_quoted = ', '.join(f"'{fn}'" for fn in files)
print(f"Found duplicates: {files_quoted}")
files.sort()
for fn in files[1:]:
print(f"Removing {fn}")
os.remove(fn)
| Remove imgur duplicates based on ID. | Remove imgur duplicates based on ID.
| Python | mit | ammongit/scripts,ammongit/scripts,ammongit/scripts,ammongit/scripts |
- import hashlib
import re
import os
import sys
- IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)')
+ IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)')
-
- def get_hash(fn):
- with open(fn, 'rb') as fh:
- hashsum = hashlib.md5(fh.read()).digest()
- return hashsum
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
- sums = {}
+ ids = {}
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
- new_fn = f'{match.group(1)}.{match.group(2)}'
+ new_fn = f'{match[1]}.{match[3]}'
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
- hashsum = get_hash(new_fn)
+ id = match[2]
- files = sums.get(hashsum, [])
+ files = ids.get(id, [])
files.append(new_fn)
- sums[hashsum] = files
+ ids[id] = files
- for hashsum, files in sums.items():
+ for _, files in ids.items():
if len(files) > 1:
- files_quoted = [f"'{x}'" for x in files]
+ files_quoted = ', '.join(f"'{fn}'" for fn in files)
- print(f"Found duplicates: {', '.join(files_quoted)}")
+ print(f"Found duplicates: {files_quoted}")
files.sort()
for fn in files[1:]:
+ print(f"Removing {fn}")
os.remove(fn)
| Remove imgur duplicates based on ID. | ## Code Before:
import hashlib
import re
import os
import sys
IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)')
def get_hash(fn):
with open(fn, 'rb') as fh:
hashsum = hashlib.md5(fh.read()).digest()
return hashsum
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
sums = {}
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
new_fn = f'{match.group(1)}.{match.group(2)}'
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
hashsum = get_hash(new_fn)
files = sums.get(hashsum, [])
files.append(new_fn)
sums[hashsum] = files
for hashsum, files in sums.items():
if len(files) > 1:
files_quoted = [f"'{x}'" for x in files]
print(f"Found duplicates: {', '.join(files_quoted)}")
files.sort()
for fn in files[1:]:
os.remove(fn)
## Instruction:
Remove imgur duplicates based on ID.
## Code After:
import re
import os
import sys
IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)')
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
ids = {}
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
new_fn = f'{match[1]}.{match[3]}'
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
id = match[2]
files = ids.get(id, [])
files.append(new_fn)
ids[id] = files
for _, files in ids.items():
if len(files) > 1:
files_quoted = ', '.join(f"'{fn}'" for fn in files)
print(f"Found duplicates: {files_quoted}")
files.sort()
for fn in files[1:]:
print(f"Removing {fn}")
os.remove(fn)
|
- import hashlib
import re
import os
import sys
- IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)')
? --- -
+ IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)')
? +
-
- def get_hash(fn):
- with open(fn, 'rb') as fh:
- hashsum = hashlib.md5(fh.read()).digest()
- return hashsum
if __name__ == '__main__':
if len(sys.argv) >= 2:
os.chdir(sys.argv[1])
- sums = {}
? ^^^
+ ids = {}
? ^^
for fn in os.listdir('.'):
match = IMGUR_FILENAME_REGEX.match(fn)
if match is None:
continue
- new_fn = f'{match.group(1)}.{match.group(2)}'
? ^^^^^^^ ^ ^^^^^^^^^
+ new_fn = f'{match[1]}.{match[3]}'
? ^ ^ ^^^
if fn == new_fn:
continue
print(f"Renaming '{fn}' to '{new_fn}'")
os.rename(fn, new_fn)
- hashsum = get_hash(new_fn)
+ id = match[2]
- files = sums.get(hashsum, [])
? ^^^ ^^^^^^^
+ files = ids.get(id, [])
? ^^ ^^
files.append(new_fn)
- sums[hashsum] = files
+ ids[id] = files
- for hashsum, files in sums.items():
? ^^^^^^^ ^^^
+ for _, files in ids.items():
? ^ ^^
if len(files) > 1:
- files_quoted = [f"'{x}'" for x in files]
? ^ ^ ^ ^
+ files_quoted = ', '.join(f"'{fn}'" for fn in files)
? ^^^^^^^^^^ ^^ ^^ ^
- print(f"Found duplicates: {', '.join(files_quoted)}")
? ---------- -
+ print(f"Found duplicates: {files_quoted}")
files.sort()
for fn in files[1:]:
+ print(f"Removing {fn}")
os.remove(fn)
|
efe1417ad049e4bb78bf1f111db6b2ea9c603461 | rapt/util.py | rapt/util.py | import sys
import yaml
import click
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip()
| import sys
import yaml
import click
def load_yaml(fh_or_string):
return yaml.safe_load(fh_or_string)
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip()
| Add a load yaml helper | Add a load yaml helper
| Python | bsd-3-clause | yougov/rapt,yougov/rapt | import sys
import yaml
import click
+
+
+ def load_yaml(fh_or_string):
+ return yaml.safe_load(fh_or_string)
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip()
| Add a load yaml helper | ## Code Before:
import sys
import yaml
import click
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip()
## Instruction:
Add a load yaml helper
## Code After:
import sys
import yaml
import click
def load_yaml(fh_or_string):
return yaml.safe_load(fh_or_string)
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip()
| import sys
import yaml
import click
+
+
+ def load_yaml(fh_or_string):
+ return yaml.safe_load(fh_or_string)
def dump_yaml(obj):
return yaml.dump(obj, default_flow_style=False)
def edit_yaml(content='', footer=''):
MARKER = '# Everything below is ignored\n\n'
message = click.edit(content + '\n\n' + MARKER + footer,
extension='.yaml')
if message is not None:
yaml_content = message.split(MARKER, 1)[0].rstrip('\n')
return yaml.safe_load(yaml_content)
def stdin():
for line in sys.stdin:
yield line.strip() |
6a15b33d69d8d66643bb8886f9916fa28ecaedea | molo/yourwords/templatetags/competition_tag.py | molo/yourwords/templatetags/competition_tag.py | from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = []
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
| from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = YourWordsCompetition.objects.none()
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
| Return None if there is no competition | Return None if there is no competition
| Python | bsd-2-clause | praekelt/molo.yourwords,praekelt/molo.yourwords | from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
- competitions = []
+ competitions = YourWordsCompetition.objects.none()
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
| Return None if there is no competition | ## Code Before:
from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = []
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
## Instruction:
Return None if there is no competition
## Code After:
from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
competitions = YourWordsCompetition.objects.none()
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return []
| from django import template
from copy import copy
from molo.yourwords.models import (YourWordsCompetition, ThankYou,
YourWordsCompetitionIndexPage)
from molo.core.templatetags.core_tags import get_pages
register = template.Library()
@register.inclusion_tag(
'yourwords/your_words_competition_tag.html',
takes_context=True
)
def your_words_competition(context):
context = copy(context)
locale_code = context.get('locale_code')
page = YourWordsCompetitionIndexPage.objects.live().all().first()
if page:
competitions = (
YourWordsCompetition.objects.child_of(page).filter(
languages__language__is_main_language=True).specific())
else:
- competitions = []
+ competitions = YourWordsCompetition.objects.none()
context.update({
'competitions': get_pages(context, competitions, locale_code)
})
return context
@register.assignment_tag(takes_context=True)
def load_thank_you_page_for_competition(context, competition):
page = competition.get_main_language_page()
locale = context.get('locale_code')
qs = ThankYou.objects.child_of(page).filter(
languages__language__is_main_language=True)
if not locale:
return qs
if qs:
return get_pages(context, qs, locale)
else:
return [] |
a42a6a54f732ca7eba700b867a3025739ad6a271 | list_all_users_in_group.py | list_all_users_in_group.py |
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
if __name__ == "__main__":
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
|
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
def main():
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
if __name__ == "__main__":
main()
| Move main code to function because of pylint warning 'Invalid constant name' | Move main code to function because of pylint warning 'Invalid constant name'
| Python | cc0-1.0 | vazhnov/list_all_users_in_group |
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
- if __name__ == "__main__":
+
+ def main():
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
+
+ if __name__ == "__main__":
+ main()
+ | Move main code to function because of pylint warning 'Invalid constant name' | ## Code Before:
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
if __name__ == "__main__":
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
## Instruction:
Move main code to function because of pylint warning 'Invalid constant name'
## Code After:
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
def main():
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
if __name__ == "__main__":
main()
|
from __future__ import print_function
import grp
import pwd
import inspect
import argparse
def list_all_users_in_group(groupname):
"""Get list of all users of group.
Get sorted list of all users of group GROUP,
including users with main group GROUP.
Origin in https://github.com/vazhnov/list_all_users_in_group
"""
try:
group = grp.getgrnam(groupname)
# On error "KeyError: 'getgrnam(): name not found: GROUP'"
except KeyError:
return None
group_all_users_set = set(group.gr_mem)
for user in pwd.getpwall():
if user.pw_gid == group.gr_gid:
group_all_users_set.add(user.pw_name)
return sorted(group_all_users_set)
- if __name__ == "__main__":
+
+ def main():
parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group),
formatter_class=argparse.RawDescriptionHelpFormatter)
parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter')
parser.add_argument('groupname', help='Group name')
args = parser.parse_args()
result = list_all_users_in_group(args.groupname)
if result:
print (args.delimiter.join(result))
+
+
+ if __name__ == "__main__":
+ main() |
4072f8ec6e1908d6e84859c8a0bd6c96562ea5cc | parts/plugins/x-shell.py | parts/plugins/x-shell.py | import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
| import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
'SNAPDIR=' + self.builddir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
# vim:set ts=4 noet:
| Add "SNAPDIR" and simple vim modeline | Add "SNAPDIR" and simple vim modeline
| Python | mit | infosiftr/snap-docker,docker-snap/docker,docker-snap/docker | import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
+ 'SNAPDIR=' + self.builddir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
+ # vim:set ts=4 noet:
+ | Add "SNAPDIR" and simple vim modeline | ## Code Before:
import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
## Instruction:
Add "SNAPDIR" and simple vim modeline
## Code After:
import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
'SNAPDIR=' + self.builddir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
# vim:set ts=4 noet:
| import snapcraft
class ShellPlugin(snapcraft.BasePlugin):
@classmethod
def schema(cls):
schema = super().schema()
schema['required'] = []
schema['properties']['shell'] = {
'type': 'string',
'default': '/bin/sh',
}
schema['required'].append('shell')
schema['properties']['shell-flags'] = {
'type': 'array',
'items': {
'type': 'string',
},
'default': [],
}
schema['properties']['shell-command'] = {
'type': 'string',
}
schema['required'].append('shell-command')
return schema
def env(self, root):
return super().env(root) + [
'DESTDIR=' + self.installdir,
+ 'SNAPDIR=' + self.builddir,
]
def build(self):
super().build()
return self.run([
self.options.shell,
] + self.options.shell_flags + [
'-c', self.options.shell_command,
])
+
+ # vim:set ts=4 noet: |
2ad2d488b4d7b0997355c068646a6a38b2668dae | meetuppizza/tests.py | meetuppizza/tests.py | from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
| from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
def test_page_contains_pizza(self):
response = self.client.get('/')
self.assertContains(response, "Pizza")
| Add test that checks if landing page contains the word Pizza. | Add test that checks if landing page contains the word Pizza.
| Python | mit | nicole-a-tesla/meetup.pizza,nicole-a-tesla/meetup.pizza | from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
+ def test_page_contains_pizza(self):
+ response = self.client.get('/')
+ self.assertContains(response, "Pizza")
+ | Add test that checks if landing page contains the word Pizza. | ## Code Before:
from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
## Instruction:
Add test that checks if landing page contains the word Pizza.
## Code After:
from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
def test_page_contains_pizza(self):
response = self.client.get('/')
self.assertContains(response, "Pizza")
| from django.test import TestCase
class Test(TestCase):
def test_landing_page_is_there(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 200)
+
+ def test_page_contains_pizza(self):
+ response = self.client.get('/')
+ self.assertContains(response, "Pizza") |
a35a25732159e4c8b5655755ce31ec4c3e6e7975 | dummy_robot/dummy_robot_bringup/launch/dummy_robot_bringup.launch.py | dummy_robot/dummy_robot_bringup/launch/dummy_robot_bringup.launch.py |
import os
from ament_index_python.packages import get_package_share_directory
from launch import LaunchDescription
from launch_ros.actions import Node
def generate_launch_description():
# TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch
urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'),
'launch', 'single_rrbot.urdf')
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
Node(package='robot_state_publisher', node_executable='robot_state_publisher',
output='screen', arguments=[urdf]),
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
])
|
import os
from launch import LaunchDescription
from launch_ros.actions import Node
from launch_ros.substitutions import FindPackageShare
def generate_launch_description():
pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup')
urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf')
with open(urdf_file, 'r') as infp:
robot_desc = infp.read()
rsp_params = {'robot_description': robot_desc}
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
Node(package='robot_state_publisher', node_executable='robot_state_publisher_node',
output='screen', parameters=[rsp_params]),
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
])
| Switch dummy_robot_bringup to use parameter for rsp. | Switch dummy_robot_bringup to use parameter for rsp.
Signed-off-by: Chris Lalancette <281cd07d7578d97c83271fbbf2faddb83ab3791c@openrobotics.org>
| Python | apache-2.0 | ros2/demos,ros2/demos,ros2/demos,ros2/demos |
import os
- from ament_index_python.packages import get_package_share_directory
from launch import LaunchDescription
from launch_ros.actions import Node
+ from launch_ros.substitutions import FindPackageShare
def generate_launch_description():
- # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch
- urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'),
- 'launch', 'single_rrbot.urdf')
+ pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup')
+ urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf')
+ with open(urdf_file, 'r') as infp:
+ robot_desc = infp.read()
+ rsp_params = {'robot_description': robot_desc}
+
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
- Node(package='robot_state_publisher', node_executable='robot_state_publisher',
+ Node(package='robot_state_publisher', node_executable='robot_state_publisher_node',
- output='screen', arguments=[urdf]),
+ output='screen', parameters=[rsp_params]),
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
])
| Switch dummy_robot_bringup to use parameter for rsp. | ## Code Before:
import os
from ament_index_python.packages import get_package_share_directory
from launch import LaunchDescription
from launch_ros.actions import Node
def generate_launch_description():
# TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch
urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'),
'launch', 'single_rrbot.urdf')
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
Node(package='robot_state_publisher', node_executable='robot_state_publisher',
output='screen', arguments=[urdf]),
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
])
## Instruction:
Switch dummy_robot_bringup to use parameter for rsp.
## Code After:
import os
from launch import LaunchDescription
from launch_ros.actions import Node
from launch_ros.substitutions import FindPackageShare
def generate_launch_description():
pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup')
urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf')
with open(urdf_file, 'r') as infp:
robot_desc = infp.read()
rsp_params = {'robot_description': robot_desc}
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
Node(package='robot_state_publisher', node_executable='robot_state_publisher_node',
output='screen', parameters=[rsp_params]),
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
])
|
import os
- from ament_index_python.packages import get_package_share_directory
from launch import LaunchDescription
from launch_ros.actions import Node
+ from launch_ros.substitutions import FindPackageShare
def generate_launch_description():
- # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch
- urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'),
- 'launch', 'single_rrbot.urdf')
+ pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup')
+ urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf')
+ with open(urdf_file, 'r') as infp:
+ robot_desc = infp.read()
+ rsp_params = {'robot_description': robot_desc}
+
return LaunchDescription([
Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'),
- Node(package='robot_state_publisher', node_executable='robot_state_publisher',
+ Node(package='robot_state_publisher', node_executable='robot_state_publisher_node',
? +++++
- output='screen', arguments=[urdf]),
? ^^ - - ^^
+ output='screen', parameters=[rsp_params]),
? + ^ ++ ^^^^^^^^^
Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'),
Node(package='dummy_sensors', node_executable='dummy_laser', output='screen')
]) |
456e5a63333e683b7167bf151b97a49a5cf5c5fe | app/models/job.py | app/models/job.py |
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
def __init__(self, name):
super(JobDocument, self).__init__(name)
@property
def collection(self):
return JOB_COLLECTION
|
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
def __init__(self, name, job=None, kernel=None):
super(JobDocument, self).__init__(name)
self._private = False
self._job = job
self._kernel = kernel
self._created = None
@property
def collection(self):
return JOB_COLLECTION
@property
def private(self):
return self._private
@private.setter
def private(self, value):
self._private = value
@property
def job(self):
return self._job
@job.setter
def job(self, value):
self._job = value
@property
def kernel(self):
return self._kernel
@kernel.setter
def kernel(self, value):
self._kernel = value
@property
def created(self):
return self._created
@created.setter
def created(self, value):
self._created = value
def to_dict(self):
job_dict = super(JobDocument, self).to_dict()
job_dict['private'] = self._private
job_dict['job'] = self._job
job_dict['kernel'] = self._kernel
job_dict['created'] = str(self._created)
return job_dict
| Rework the jod document model. | Rework the jod document model.
* Add the created field that will store a datetime object.
* Add reference to the kernel and the job inside the document,
without relying on the Jod document name itself. Since we
use the dash as a separator, and other job names can have
dash in them, we cannot separate job from kernel easily.
| Python | agpl-3.0 | joyxu/kernelci-backend,joyxu/kernelci-backend,kernelci/kernelci-backend,kernelci/kernelci-backend,joyxu/kernelci-backend |
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
- def __init__(self, name):
+ def __init__(self, name, job=None, kernel=None):
super(JobDocument, self).__init__(name)
+
+ self._private = False
+ self._job = job
+ self._kernel = kernel
+ self._created = None
@property
def collection(self):
return JOB_COLLECTION
+ @property
+ def private(self):
+ return self._private
+
+ @private.setter
+ def private(self, value):
+ self._private = value
+
+ @property
+ def job(self):
+ return self._job
+
+ @job.setter
+ def job(self, value):
+ self._job = value
+
+ @property
+ def kernel(self):
+ return self._kernel
+
+ @kernel.setter
+ def kernel(self, value):
+ self._kernel = value
+
+ @property
+ def created(self):
+ return self._created
+
+ @created.setter
+ def created(self, value):
+ self._created = value
+
+ def to_dict(self):
+ job_dict = super(JobDocument, self).to_dict()
+ job_dict['private'] = self._private
+ job_dict['job'] = self._job
+ job_dict['kernel'] = self._kernel
+ job_dict['created'] = str(self._created)
+ return job_dict
+ | Rework the jod document model. | ## Code Before:
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
def __init__(self, name):
super(JobDocument, self).__init__(name)
@property
def collection(self):
return JOB_COLLECTION
## Instruction:
Rework the jod document model.
## Code After:
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
def __init__(self, name, job=None, kernel=None):
super(JobDocument, self).__init__(name)
self._private = False
self._job = job
self._kernel = kernel
self._created = None
@property
def collection(self):
return JOB_COLLECTION
@property
def private(self):
return self._private
@private.setter
def private(self, value):
self._private = value
@property
def job(self):
return self._job
@job.setter
def job(self, value):
self._job = value
@property
def kernel(self):
return self._kernel
@kernel.setter
def kernel(self, value):
self._kernel = value
@property
def created(self):
return self._created
@created.setter
def created(self, value):
self._created = value
def to_dict(self):
job_dict = super(JobDocument, self).to_dict()
job_dict['private'] = self._private
job_dict['job'] = self._job
job_dict['kernel'] = self._kernel
job_dict['created'] = str(self._created)
return job_dict
|
from base import BaseDocument
JOB_COLLECTION = 'job'
class JobDocument(BaseDocument):
JOB_ID_FORMAT = '%s-%s'
- def __init__(self, name):
+ def __init__(self, name, job=None, kernel=None):
super(JobDocument, self).__init__(name)
+
+ self._private = False
+ self._job = job
+ self._kernel = kernel
+ self._created = None
@property
def collection(self):
return JOB_COLLECTION
+
+ @property
+ def private(self):
+ return self._private
+
+ @private.setter
+ def private(self, value):
+ self._private = value
+
+ @property
+ def job(self):
+ return self._job
+
+ @job.setter
+ def job(self, value):
+ self._job = value
+
+ @property
+ def kernel(self):
+ return self._kernel
+
+ @kernel.setter
+ def kernel(self, value):
+ self._kernel = value
+
+ @property
+ def created(self):
+ return self._created
+
+ @created.setter
+ def created(self, value):
+ self._created = value
+
+ def to_dict(self):
+ job_dict = super(JobDocument, self).to_dict()
+ job_dict['private'] = self._private
+ job_dict['job'] = self._job
+ job_dict['kernel'] = self._kernel
+ job_dict['created'] = str(self._created)
+ return job_dict |
ab802204d84511765a701cad48e9e22dc4e84be1 | tests/rules/conftest.py | tests/rules/conftest.py | import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True, scope="session")
def configured_cache():
cache.configure()
| import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True)
def configured_cache():
if not cache.region.is_configured:
cache.configure()
yield
cache.region.invalidate()
| Fix intermittent failures of test_guard_http_exception | Fix intermittent failures of test_guard_http_exception
Signed-off-by: Ryan Lerch <e809e25f3c554b2b195ccd768cd9a485288f896f@redhat.com>
| Python | lgpl-2.1 | fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn | import pytest
from fmn.rules.cache import cache
- @pytest.fixture(autouse=True, scope="session")
+ @pytest.fixture(autouse=True)
def configured_cache():
+ if not cache.region.is_configured:
- cache.configure()
+ cache.configure()
+ yield
+ cache.region.invalidate()
| Fix intermittent failures of test_guard_http_exception | ## Code Before:
import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True, scope="session")
def configured_cache():
cache.configure()
## Instruction:
Fix intermittent failures of test_guard_http_exception
## Code After:
import pytest
from fmn.rules.cache import cache
@pytest.fixture(autouse=True)
def configured_cache():
if not cache.region.is_configured:
cache.configure()
yield
cache.region.invalidate()
| import pytest
from fmn.rules.cache import cache
- @pytest.fixture(autouse=True, scope="session")
? -----------------
+ @pytest.fixture(autouse=True)
def configured_cache():
+ if not cache.region.is_configured:
- cache.configure()
+ cache.configure()
? ++++
+ yield
+ cache.region.invalidate() |
cfb50f4ff62770c397634897e09497b74b396067 | notifications/level_starting.py | notifications/level_starting.py | from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data
| from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
data['message_data']['event_key'] = self.event.key_name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data
| Add event key to comp level starting notification | Add event key to comp level starting notification
| Python | mit | josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,phil-lopreiato/the-blue-alliance,phil-lopreiato/the-blue-alliance,fangeugene/the-blue-alliance,bvisness/the-blue-alliance,nwalters512/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,the-blue-alliance/the-blue-alliance,verycumbersome/the-blue-alliance,bdaroz/the-blue-alliance,bvisness/the-blue-alliance,jaredhasenklein/the-blue-alliance,fangeugene/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,nwalters512/the-blue-alliance,the-blue-alliance/the-blue-alliance,jaredhasenklein/the-blue-alliance,phil-lopreiato/the-blue-alliance,the-blue-alliance/the-blue-alliance,the-blue-alliance/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,bdaroz/the-blue-alliance,verycumbersome/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,1fish2/the-blue-alliance,1fish2/the-blue-alliance,tsteward/the-blue-alliance,verycumbersome/the-blue-alliance,1fish2/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,bvisness/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,jaredhasenklein/the-blue-alliance,bdaroz/the-blue-alliance,tsteward/the-blue-alliance,tsteward/the-blue-alliance,nwalters512/the-blue-alliance,synth3tk/the-blue-alliance,synth3tk/the-blue-alliance,bdaroz/the-blue-alliance,synth3tk/the-blue-alliance,tsteward/the-blue-alliance,1fish2/the-blue-alliance,1fish2/the-blue-alliance,verycumbersome/the-blue-alliance,verycumbersome/the-blue-alliance,tsteward/the-blue-alliance,synth3tk/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,the-blue-alliance/the-blue-alliance,phil-lopreiato/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,1fish2/the-blue-alliance,the-blue-alliance/the-blue-alliance,synth3tk/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance | from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
+ data['message_data']['event_key'] = self.event.key_name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data
| Add event key to comp level starting notification | ## Code Before:
from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data
## Instruction:
Add event key to comp level starting notification
## Code After:
from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
data['message_data']['event_key'] = self.event.key_name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data
| from consts.notification_type import NotificationType
from notifications.base_notification import BaseNotification
class CompLevelStartingNotification(BaseNotification):
def __init__(self, match, event):
self.match = match
self.event = event
def _build_dict(self):
data = {}
data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING]
data['message_data'] = {}
data['message_data']['event_name'] = self.event.name
+ data['message_data']['event_key'] = self.event.key_name
data['message_data']['comp_level'] = self.match.comp_level
data['message_data']['scheduled_time'] = self.match.time
return data |
b4578d34adaa641dab5082f9d2bffe14c69649c5 | detour/__init__.py | detour/__init__.py | from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
| from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
class DetourException(NotImplementedError):
pass
| Add a root exception for use if necessary. | Add a root exception for use if necessary.
| Python | bsd-2-clause | kezabelle/wsgi-detour | from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
+
+ class DetourException(NotImplementedError):
+ pass
+ | Add a root exception for use if necessary. | ## Code Before:
from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
## Instruction:
Add a root exception for use if necessary.
## Code After:
from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
class DetourException(NotImplementedError):
pass
| from __future__ import absolute_import
from __future__ import unicode_literals
__version_info__ = '0.1.0'
__version__ = '0.1.0'
version = '0.1.0'
VERSION = '0.1.0'
def get_version():
return version # pragma: no cover
+
+
+ class DetourException(NotImplementedError):
+ pass |
55a1f6197800249b3ad13ec7c5358e907ea04c46 | comics/comics/treadingground.py | comics/comics/treadingground.py | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
history_capable_days = 30
schedule = 'Mo,We,Fr'
time_zone = -5
def crawl(self, pub_date):
pass
| from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
schedule = None
def crawl(self, pub_date):
pass # Comic no longer published
| Remove schedule for ended comic | Remove schedule for ended comic
| Python | agpl-3.0 | klette/comics,datagutten/comics,klette/comics,jodal/comics,datagutten/comics,klette/comics,jodal/comics,jodal/comics,jodal/comics,datagutten/comics,datagutten/comics | from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
- history_capable_days = 30
- schedule = 'Mo,We,Fr'
+ schedule = None
- time_zone = -5
def crawl(self, pub_date):
- pass
+ pass # Comic no longer published
| Remove schedule for ended comic | ## Code Before:
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
history_capable_days = 30
schedule = 'Mo,We,Fr'
time_zone = -5
def crawl(self, pub_date):
pass
## Instruction:
Remove schedule for ended comic
## Code After:
from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
schedule = None
def crawl(self, pub_date):
pass # Comic no longer published
| from comics.aggregator.crawler import CrawlerBase, CrawlerImage
from comics.meta.base import MetaBase
class Meta(MetaBase):
name = 'Treading Ground'
language = 'en'
url = 'http://www.treadingground.com/'
start_date = '2003-10-12'
rights = 'Nick Wright'
class Crawler(CrawlerBase):
- history_capable_days = 30
- schedule = 'Mo,We,Fr'
? ^^ ^^ ----
+ schedule = None
? ^ ^
- time_zone = -5
def crawl(self, pub_date):
- pass
+ pass # Comic no longer published |
1bc4507234d87b1ed246501165fa1d8138bf5ca6 | cheddar/exceptions.py | cheddar/exceptions.py |
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
|
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
| Fix for pypy compatibility: must super's __init__ | Fix for pypy compatibility: must super's __init__
| Python | apache-2.0 | jessemyers/cheddar,jessemyers/cheddar |
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
+ super(NotFoundError, self).__init__()
self.status_code = status_code
| Fix for pypy compatibility: must super's __init__ | ## Code Before:
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
## Instruction:
Fix for pypy compatibility: must super's __init__
## Code After:
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
+ super(NotFoundError, self).__init__()
self.status_code = status_code
|
e6c43333c3939247534ddee4c419dcdcff5eda5f | spyder_terminal/server/rest/term_rest.py | spyder_terminal/server/rest/term_rest.py |
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
rows = int(self.get_argument('rows', None, 23))
cols = int(self.get_argument('cols', None, 73))
self.application.term_manager.resize_term(pid, rows, cols)
|
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
self.application.term_manager.resize_term(pid, rows, cols)
| Change default terminal resize arguments | Change default terminal resize arguments
| Python | mit | andfoy/spyder-terminal,spyder-ide/spyder-terminal,spyder-ide/spyder-terminal,andfoy/spyder-terminal,andfoy/spyder-terminal,spyder-ide/spyder-terminal,spyder-ide/spyder-terminal |
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
- rows = int(self.get_argument('rows', None, 23))
+ rows = int(self.get_argument('rows', default=23))
- cols = int(self.get_argument('cols', None, 73))
+ cols = int(self.get_argument('cols', default=73))
self.application.term_manager.resize_term(pid, rows, cols)
| Change default terminal resize arguments | ## Code Before:
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
rows = int(self.get_argument('rows', None, 23))
cols = int(self.get_argument('cols', None, 73))
self.application.term_manager.resize_term(pid, rows, cols)
## Instruction:
Change default terminal resize arguments
## Code After:
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
self.application.term_manager.resize_term(pid, rows, cols)
|
"""Main HTTP routes request handlers."""
import tornado.web
import tornado.escape
from os import getcwd
class MainHandler(tornado.web.RequestHandler):
"""Handles creation of new terminals."""
@tornado.gen.coroutine
def post(self):
"""POST verb: Create a new terminal."""
rows = int(self.get_argument('rows', default=23))
cols = int(self.get_argument('cols', default=73))
cwd = self.get_cookie('cwd', default=getcwd())
self.application.logger.info('CWD: {0}'.format(cwd))
self.application.logger.info('Size: ({0}, {1})'.format(cols, rows))
pid = yield self.application.term_manager.create_term(rows, cols, cwd)
self.write(pid)
class ResizeHandler(tornado.web.RequestHandler):
"""Handles resizing of terminals."""
@tornado.gen.coroutine
def post(self, pid):
"""POST verb: Resize a terminal."""
- rows = int(self.get_argument('rows', None, 23))
? ^^^ ^^
+ rows = int(self.get_argument('rows', default=23))
? ^ ^^^^^^
- cols = int(self.get_argument('cols', None, 73))
? ^^^ ^^
+ cols = int(self.get_argument('cols', default=73))
? ^ ^^^^^^
self.application.term_manager.resize_term(pid, rows, cols) |
746510dc0b939fe11a2b025805678a0829cf814a | handler/minion_server.py | handler/minion_server.py |
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main() |
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
"""Start a worker.
Message format:
{
'image': 'image name'
'numprocs': number of workers,
'args': 'extra arguments for "docker run -d image ..."'
}
"""
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main() | Document message format for minion server | Document message format for minion server
| Python | mit | waltermoreira/adama-minion |
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
+ """Start a worker.
+
+ Message format:
+
+ {
+ 'image': 'image name'
+ 'numprocs': number of workers,
+ 'args': 'extra arguments for "docker run -d image ..."'
+ }
+
+ """
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main() | Document message format for minion server | ## Code Before:
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main()
## Instruction:
Document message format for minion server
## Code After:
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
"""Start a worker.
Message format:
{
'image': 'image name'
'numprocs': number of workers,
'args': 'extra arguments for "docker run -d image ..."'
}
"""
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main() |
import server
import supervisor
class MinionServer(server.Server):
def __init__(self, ip, port):
super(MinionServer, self).__init__(ip, port)
def handle(self, data):
+ """Start a worker.
+
+ Message format:
+
+ {
+ 'image': 'image name'
+ 'numprocs': number of workers,
+ 'args': 'extra arguments for "docker run -d image ..."'
+ }
+
+ """
supervisor.start(
'worker.conf',
target='worker_{}'.format(data['image']),
image=data['image'],
numprocs=data.get('numprocs', 1),
args=data.get('args', ''))
return {'status': 'ok'}
def main():
server = MinionServer('*', 1234)
server.start()
server.join()
if __name__ == '__main__':
main() |
414dd0b03b3e4eabc11f848f79d681f3a284380e | pygcvs/helpers.py | pygcvs/helpers.py | from .parser import GcvsParser
try:
import ephem
except ImportError:
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
if ephem is None:
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body
| from .parser import GcvsParser
try:
import ephem
except ImportError: # pragma: no cover
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
if ephem is None: # pragma: no cover
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body
| Exclude missing ephem from coverage | Exclude missing ephem from coverage
| Python | mit | zsiciarz/pygcvs | from .parser import GcvsParser
try:
import ephem
- except ImportError:
+ except ImportError: # pragma: no cover
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
- if ephem is None:
+ if ephem is None: # pragma: no cover
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body
| Exclude missing ephem from coverage | ## Code Before:
from .parser import GcvsParser
try:
import ephem
except ImportError:
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
if ephem is None:
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body
## Instruction:
Exclude missing ephem from coverage
## Code After:
from .parser import GcvsParser
try:
import ephem
except ImportError: # pragma: no cover
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
if ephem is None: # pragma: no cover
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body
| from .parser import GcvsParser
try:
import ephem
- except ImportError:
+ except ImportError: # pragma: no cover
ephem = None
def read_gcvs(filename):
"""
Reads variable star data in `GCVS format`_.
:param filename: path to GCVS data file (usually ``iii.dat``)
.. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/
"""
with open(filename, 'r') as fp:
parser = GcvsParser(fp)
for star in parser:
yield star
def dict_to_body(star_dict):
"""
Converts a dictionary of variable star data to a `Body` instance.
Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed.
"""
- if ephem is None:
+ if ephem is None: # pragma: no cover
raise NotImplementedError("Please install PyEphem in order to use dict_to_body.")
body = ephem.FixedBody()
body.name = star_dict['name']
body._ra = ephem.hours(str(star_dict['ra']))
body._dec = ephem.degrees(str(star_dict['dec']))
body._epoch = ephem.J2000
return body |
4949b1051656566ce544a8240b0328a61259868a | migrations/versions/139_add_ns_index_to_contact_and_event.py | migrations/versions/139_add_ns_index_to_contact_and_event.py |
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
down_revision = '2d8a350b4885'
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.")
|
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
down_revision = '5305d4ae30b4'
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.")
| Fix migration history bug introduced with merge | Fix migration history bug introduced with merge
| Python | agpl-3.0 | wakermahmud/sync-engine,EthanBlackburn/sync-engine,Eagles2F/sync-engine,EthanBlackburn/sync-engine,gale320/sync-engine,wakermahmud/sync-engine,nylas/sync-engine,jobscore/sync-engine,nylas/sync-engine,jobscore/sync-engine,closeio/nylas,gale320/sync-engine,wakermahmud/sync-engine,closeio/nylas,PriviPK/privipk-sync-engine,PriviPK/privipk-sync-engine,jobscore/sync-engine,nylas/sync-engine,closeio/nylas,PriviPK/privipk-sync-engine,jobscore/sync-engine,ErinCall/sync-engine,Eagles2F/sync-engine,PriviPK/privipk-sync-engine,nylas/sync-engine,Eagles2F/sync-engine,EthanBlackburn/sync-engine,ErinCall/sync-engine,wakermahmud/sync-engine,wakermahmud/sync-engine,Eagles2F/sync-engine,ErinCall/sync-engine,closeio/nylas,gale320/sync-engine,PriviPK/privipk-sync-engine,gale320/sync-engine,Eagles2F/sync-engine,ErinCall/sync-engine,EthanBlackburn/sync-engine,EthanBlackburn/sync-engine,ErinCall/sync-engine,gale320/sync-engine |
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
- down_revision = '2d8a350b4885'
+ down_revision = '5305d4ae30b4'
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.")
| Fix migration history bug introduced with merge | ## Code Before:
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
down_revision = '2d8a350b4885'
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.")
## Instruction:
Fix migration history bug introduced with merge
## Code After:
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
down_revision = '5305d4ae30b4'
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.")
|
# revision identifiers, used by Alembic.
revision = '1fd7b3e0b662'
- down_revision = '2d8a350b4885'
? ^ ^ - ---
+ down_revision = '5305d4ae30b4'
? ^^^^ ^ +
from alembic import op
def upgrade():
op.create_index(
'ix_contact_ns_uid_provider_name',
'contact',
['namespace_id', 'uid', 'provider_name'], unique=False)
op.create_index(
'ix_event_ns_uid_provider_name',
'event',
['namespace_id', 'uid', 'provider_name'], unique=False)
def downgrade():
raise Exception("Don't bother.") |
ab8fc00a7dc6618d23e06f06e125da5ee69b2dba | event_registration_hr_contract/__openerp__.py | event_registration_hr_contract/__openerp__.py | {
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
'hr_contract_stages'
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
}
| {
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
}
| Remove the dependence with the module hr_contract_stage. | [IMP] event_registration_hr_contract: Remove the dependence with the module hr_contract_stage.
| Python | agpl-3.0 | avanzosc/event-wip | {
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
- 'hr_contract_stages'
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
}
| Remove the dependence with the module hr_contract_stage. | ## Code Before:
{
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
'hr_contract_stages'
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
}
## Instruction:
Remove the dependence with the module hr_contract_stage.
## Code After:
{
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
}
| {
"name": "Event Registration Hr Contract",
'version': '8.0.1.1.0',
'license': "AGPL-3",
'author': "AvanzOSC",
'website': "http://www.avanzosc.es",
'contributors': [
"Ana Juaristi <anajuaristi@avanzosc.es>",
"Alfredo de la Fuente <alfredodelafuente@avanzosc.es",
],
"category": "Event Management",
"depends": [
'event_track_presence_hr_holidays',
- 'hr_contract_stages'
],
"data": [
'wizard/wiz_calculate_employee_calendar_view.xml',
'wizard/wiz_event_append_assistant_view.xml',
'views/event_event_view.xml',
'views/hr_contract_view.xml',
'views/event_track_presence_view.xml',
'views/res_partner_calendar_view.xml',
'views/res_partner_calendar_day_view.xml'
],
"installable": True,
} |
6795e02c14fa99da2c0812fe6694bbd503f89ad1 | tests/mock_vws/test_invalid_given_id.py | tests/mock_vws/test_invalid_given_id.py |
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
endpoint: TargetAPIEndpoint,
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
)
|
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
any_endpoint: TargetAPIEndpoint,
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
endpoint = any_endpoint
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
)
| Use any_endpoint on invalid id test | Use any_endpoint on invalid id test
| Python | mit | adamtheturtle/vws-python,adamtheturtle/vws-python |
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
- endpoint: TargetAPIEndpoint,
+ any_endpoint: TargetAPIEndpoint,
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
+ endpoint = any_endpoint
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
)
| Use any_endpoint on invalid id test | ## Code Before:
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
endpoint: TargetAPIEndpoint,
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
)
## Instruction:
Use any_endpoint on invalid id test
## Code After:
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
any_endpoint: TargetAPIEndpoint,
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
endpoint = any_endpoint
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
)
|
import pytest
import requests
from requests import codes
from mock_vws._constants import ResultCodes
from tests.mock_vws.utils import (
TargetAPIEndpoint,
VuforiaDatabaseKeys,
assert_vws_failure,
delete_target,
)
@pytest.mark.usefixtures('verify_mock_vuforia')
class TestInvalidGivenID:
"""
Tests for giving an invalid ID to endpoints which require a target ID to
be given.
"""
def test_not_real_id(
self,
vuforia_database_keys: VuforiaDatabaseKeys,
- endpoint: TargetAPIEndpoint,
+ any_endpoint: TargetAPIEndpoint,
? ++++
target_id: str,
) -> None:
"""
A `NOT_FOUND` error is returned when an endpoint is given a target ID
of a target which does not exist.
"""
+ endpoint = any_endpoint
if not endpoint.prepared_request.path_url.endswith(target_id):
return
delete_target(
vuforia_database_keys=vuforia_database_keys,
target_id=target_id,
)
session = requests.Session()
response = session.send( # type: ignore
request=endpoint.prepared_request,
)
assert_vws_failure(
response=response,
status_code=codes.NOT_FOUND,
result_code=ResultCodes.UNKNOWN_TARGET,
) |
de42731ab97a7d4272c44cc750891906aa5b4417 | buildlet/runner/ipythonparallel.py | buildlet/runner/ipythonparallel.py |
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
self.view.wait(self.results.values())
|
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
for r in self.results.values():
r.get()
| Raise error if any in IPythonParallelRunner.wait_tasks | Raise error if any in IPythonParallelRunner.wait_tasks
| Python | bsd-3-clause | tkf/buildlet |
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
- self.view.wait(self.results.values())
+ for r in self.results.values():
+ r.get()
| Raise error if any in IPythonParallelRunner.wait_tasks | ## Code Before:
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
self.view.wait(self.results.values())
## Instruction:
Raise error if any in IPythonParallelRunner.wait_tasks
## Code After:
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
for r in self.results.values():
r.get()
|
import IPython.parallel
from .simple import SimpleRunner
from .mixinparallel import MixInParallelRunner
class IPythonParallelRunner(MixInParallelRunner, SimpleRunner):
def submit_tasks(self):
self.client = IPython.parallel.Client()
self.view = view = self.client.load_balanced_view()
self.results = results = {}
for node in self.sorted_nodes():
deps = [results[n] for n in self.graph.predecessors(node)]
with view.temp_flags(after=deps):
results[node] = view.apply_async(self.run_func,
self.nodetaskmap[node])
def wait_tasks(self):
- self.view.wait(self.results.values())
+ for r in self.results.values():
+ r.get() |
d565fdab9cefc080ff3127f036c19e95cba73f6e | tests/test_udacity.py | tests/test_udacity.py | import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
def test_udacity_api_mongofy_courses(self):
course = self.udacity_test_object.mongofy_courses()[0]
self.assertEqual(course['title'], 'Intro to Computer Science')
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| Add unit test for mongofy_courses of udacity module | Add unit test for mongofy_courses of udacity module
| Python | mit | ueg1990/mooc_aggregator_restful_api | import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
+ def test_udacity_api_mongofy_courses(self):
+ course = self.udacity_test_object.mongofy_courses()[0]
+ self.assertEqual(course['title'], 'Intro to Computer Science')
+
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| Add unit test for mongofy_courses of udacity module | ## Code Before:
import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
## Instruction:
Add unit test for mongofy_courses of udacity module
## Code After:
import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
def test_udacity_api_mongofy_courses(self):
course = self.udacity_test_object.mongofy_courses()[0]
self.assertEqual(course['title'], 'Intro to Computer Science')
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main()
| import unittest
from mooc_aggregator_restful_api import udacity
class UdacityTestCase(unittest.TestCase):
'''
Unit Tests for module udacity
'''
def setUp(self):
self.udacity_test_object = udacity.UdacityAPI()
def test_udacity_api_response(self):
self.assertEqual(self.udacity_test_object.status_code(), 200)
+ def test_udacity_api_mongofy_courses(self):
+ course = self.udacity_test_object.mongofy_courses()[0]
+ self.assertEqual(course['title'], 'Intro to Computer Science')
+
def tearDown(self):
pass
if __name__ == '__main__':
unittest.main() |
d7f43a15a2e4535728e7ec5d3cb550af3eed1590 | h2o-py/h2o/tree/__init__.py | h2o-py/h2o/tree/__init__.py | from .tree import H2OTree
from .tree import H2ONode
__all__ = ["H2OTree", "H2ONode"] | from .tree import H2OTree
from .tree import H2ONode
from .tree import H2OSplitNode
from .tree import H2OLeafNode
__all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
| Include H2OSplitNode & H2OLeafNode in __all__ | Include H2OSplitNode & H2OLeafNode in __all__ | Python | apache-2.0 | michalkurka/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,h2oai/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,michalkurka/h2o-3,michalkurka/h2o-3,h2oai/h2o-3 | from .tree import H2OTree
from .tree import H2ONode
+ from .tree import H2OSplitNode
+ from .tree import H2OLeafNode
- __all__ = ["H2OTree", "H2ONode"]
+ __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
+ | Include H2OSplitNode & H2OLeafNode in __all__ | ## Code Before:
from .tree import H2OTree
from .tree import H2ONode
__all__ = ["H2OTree", "H2ONode"]
## Instruction:
Include H2OSplitNode & H2OLeafNode in __all__
## Code After:
from .tree import H2OTree
from .tree import H2ONode
from .tree import H2OSplitNode
from .tree import H2OLeafNode
__all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
| from .tree import H2OTree
from .tree import H2ONode
+ from .tree import H2OSplitNode
+ from .tree import H2OLeafNode
- __all__ = ["H2OTree", "H2ONode"]
+ __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"] |
7e6dc283dbecf4bf9674559198b4a2c06e9f4c2e | spacy/tests/regression/test_issue1799.py | spacy/tests/regression/test_issue1799.py | '''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1
| '''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
from __future__ import unicode_literals
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1
| Fix unicode import in test | Fix unicode import in test
| Python | mit | aikramer2/spaCy,aikramer2/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,explosion/spaCy,spacy-io/spaCy,recognai/spaCy,honnibal/spaCy,spacy-io/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,aikramer2/spaCy,spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,recognai/spaCy,explosion/spaCy,recognai/spaCy,honnibal/spaCy,spacy-io/spaCy | '''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
+ from __future__ import unicode_literals
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1
| Fix unicode import in test | ## Code Before:
'''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1
## Instruction:
Fix unicode import in test
## Code After:
'''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
from __future__ import unicode_literals
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1
| '''Test sentence boundaries are deserialized correctly,
even for non-projective sentences.'''
+ from __future__ import unicode_literals
import pytest
import numpy
from ... tokens import Doc
from ... vocab import Vocab
from ... attrs import HEAD, DEP
def test_issue1799():
problem_sentence = 'Just what I was looking for.'
heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402],
[0, 8206900633647566924], [18446744073709551615, 440],
[18446744073709551614, 442]], dtype='uint64')
doc = Doc(Vocab(), words='Just what I was looking for .'.split())
doc.vocab.strings.add('ROOT')
doc = doc.from_array([HEAD, DEP], heads_deps)
assert len(list(doc.sents)) == 1 |
8a080a94300403487dce023eec8467832af8ae79 | tests/core/migrations/0004_bookwithchapters.py | tests/core/migrations/0004_bookwithchapters.py | from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
if VERSION >= (1, 8):
from django.contrib.postgres.fields import ArrayField
chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
else:
chapters_field = models.Field() # Dummy field
class PostgresOnlyCreateModel(migrations.CreateModel):
def database_forwards(self, app_label, schema_editor, from_state, to_state):
if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state)
def database_backwards(self, app_label, schema_editor, from_state, to_state):
if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state)
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
operations = [
PostgresOnlyCreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
| from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
can_use_arrayfield = False
chapters_field = models.Field() # Dummy field
if VERSION >= (1, 8):
try:
from django.contrib.postgres.fields import ArrayField
chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
can_use_arrayfield = True
except ImportError:
# We can't use ArrayField if psycopg2 is not installed
pass
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
operations = []
pg_only_operations = [
migrations.CreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
def apply(self, project_state, schema_editor, collect_sql=False):
if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"):
self.operations = self.operations + self.pg_only_operations
return super(Migration, self).apply(project_state, schema_editor, collect_sql)
| Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed | Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed
| Python | bsd-2-clause | brillgen/django-import-export,copperleaftech/django-import-export,bmihelac/django-import-export,jnns/django-import-export,copperleaftech/django-import-export,jnns/django-import-export,brillgen/django-import-export,PetrDlouhy/django-import-export,jnns/django-import-export,PetrDlouhy/django-import-export,brillgen/django-import-export,django-import-export/django-import-export,bmihelac/django-import-export,django-import-export/django-import-export,PetrDlouhy/django-import-export,copperleaftech/django-import-export,daniell/django-import-export,bmihelac/django-import-export,PetrDlouhy/django-import-export,daniell/django-import-export,copperleaftech/django-import-export,bmihelac/django-import-export,jnns/django-import-export,brillgen/django-import-export,django-import-export/django-import-export,daniell/django-import-export,django-import-export/django-import-export,daniell/django-import-export | from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
+
+ can_use_arrayfield = False
+ chapters_field = models.Field() # Dummy field
if VERSION >= (1, 8):
+ try:
- from django.contrib.postgres.fields import ArrayField
+ from django.contrib.postgres.fields import ArrayField
- chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
+ chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
+ can_use_arrayfield = True
+ except ImportError:
+ # We can't use ArrayField if psycopg2 is not installed
+ pass
- else:
- chapters_field = models.Field() # Dummy field
-
-
- class PostgresOnlyCreateModel(migrations.CreateModel):
- def database_forwards(self, app_label, schema_editor, from_state, to_state):
- if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
- super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state)
-
- def database_backwards(self, app_label, schema_editor, from_state, to_state):
- if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
- super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state)
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
- operations = [
+ operations = []
- PostgresOnlyCreateModel(
+
+ pg_only_operations = [
+ migrations.CreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
+ def apply(self, project_state, schema_editor, collect_sql=False):
+ if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"):
+ self.operations = self.operations + self.pg_only_operations
+ return super(Migration, self).apply(project_state, schema_editor, collect_sql)
+ | Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed | ## Code Before:
from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
if VERSION >= (1, 8):
from django.contrib.postgres.fields import ArrayField
chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
else:
chapters_field = models.Field() # Dummy field
class PostgresOnlyCreateModel(migrations.CreateModel):
def database_forwards(self, app_label, schema_editor, from_state, to_state):
if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state)
def database_backwards(self, app_label, schema_editor, from_state, to_state):
if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state)
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
operations = [
PostgresOnlyCreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
## Instruction:
Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed
## Code After:
from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
can_use_arrayfield = False
chapters_field = models.Field() # Dummy field
if VERSION >= (1, 8):
try:
from django.contrib.postgres.fields import ArrayField
chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
can_use_arrayfield = True
except ImportError:
# We can't use ArrayField if psycopg2 is not installed
pass
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
operations = []
pg_only_operations = [
migrations.CreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
def apply(self, project_state, schema_editor, collect_sql=False):
if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"):
self.operations = self.operations + self.pg_only_operations
return super(Migration, self).apply(project_state, schema_editor, collect_sql)
| from __future__ import unicode_literals
from django import VERSION
from django.db import migrations, models
+
+ can_use_arrayfield = False
+ chapters_field = models.Field() # Dummy field
if VERSION >= (1, 8):
+ try:
- from django.contrib.postgres.fields import ArrayField
+ from django.contrib.postgres.fields import ArrayField
? ++++
- chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
+ chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None)
? ++++
+ can_use_arrayfield = True
+ except ImportError:
+ # We can't use ArrayField if psycopg2 is not installed
+ pass
- else:
- chapters_field = models.Field() # Dummy field
-
-
- class PostgresOnlyCreateModel(migrations.CreateModel):
- def database_forwards(self, app_label, schema_editor, from_state, to_state):
- if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
- super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state)
-
- def database_backwards(self, app_label, schema_editor, from_state, to_state):
- if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"):
- super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state)
class Migration(migrations.Migration):
dependencies = [
('core', '0003_withfloatfield'),
]
- operations = [
+ operations = []
? +
- PostgresOnlyCreateModel(
+
+ pg_only_operations = [
+ migrations.CreateModel(
name='BookWithChapters',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('name', models.CharField(max_length=100, verbose_name='Book name')),
('chapters', chapters_field)
],
),
]
+
+ def apply(self, project_state, schema_editor, collect_sql=False):
+ if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"):
+ self.operations = self.operations + self.pg_only_operations
+ return super(Migration, self).apply(project_state, schema_editor, collect_sql) |
71088ebbed3f6060def0455814036185c70ba194 | shopify_auth/context_processors.py | shopify_auth/context_processors.py | import shopify
def current_shop(request):
if not shopify.ShopifyResource.site:
return {'current_shop': None}
return {'current_shop': shopify.Shop.current()} | from django.conf import settings
import shopify
def shopify_context(request):
return {
'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None,
'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY,
} | Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information. | Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information. | Python | mit | funkybob/django-shopify-auth,RafaAguilar/django-shopify-auth,discolabs/django-shopify-auth,RafaAguilar/django-shopify-auth,discolabs/django-shopify-auth,funkybob/django-shopify-auth | + from django.conf import settings
import shopify
- def current_shop(request):
- if not shopify.ShopifyResource.site:
- return {'current_shop': None}
- return {'current_shop': shopify.Shop.current()}
+ def shopify_context(request):
+ return {
+ 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None,
+ 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY,
+ } | Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information. | ## Code Before:
import shopify
def current_shop(request):
if not shopify.ShopifyResource.site:
return {'current_shop': None}
return {'current_shop': shopify.Shop.current()}
## Instruction:
Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information.
## Code After:
from django.conf import settings
import shopify
def shopify_context(request):
return {
'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None,
'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY,
} | + from django.conf import settings
import shopify
- def current_shop(request):
- if not shopify.ShopifyResource.site:
- return {'current_shop': None}
- return {'current_shop': shopify.Shop.current()}
+ def shopify_context(request):
+ return {
+ 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None,
+ 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY,
+ } |
700fa0144c5276d8e31c01a243340f6cbac07e8f | sentry/client/handlers.py | sentry/client/handlers.py | import logging
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
get_client().create_from_record(record) | import logging
import sys
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
# Avoid typical config issues by overriding loggers behavior
if record.name == 'sentry.errors':
print >> sys.stderr, record.message
return
get_client().create_from_record(record) | Add a safety net for recursive logging | Add a safety net for recursive logging
| Python | bsd-3-clause | ewdurbin/sentry,hongliang5623/sentry,Photonomie/raven-python,ewdurbin/raven-python,daevaorn/sentry,NickPresta/sentry,dcramer/sentry-old,looker/sentry,NickPresta/sentry,chayapan/django-sentry,Kryz/sentry,BuildingLink/sentry,gencer/sentry,boneyao/sentry,danriti/raven-python,imankulov/sentry,nikolas/raven-python,mvaled/sentry,gencer/sentry,inspirehep/raven-python,arthurlogilab/raven-python,daevaorn/sentry,recht/raven-python,icereval/raven-python,drcapulet/sentry,ewdurbin/sentry,akheron/raven-python,boneyao/sentry,JamesMura/sentry,zenefits/sentry,johansteffner/raven-python,ewdurbin/sentry,gg7/sentry,SilentCircle/sentry,dbravender/raven-python,alex/raven,alex/sentry,zenefits/sentry,someonehan/raven-python,JTCunning/sentry,JTCunning/sentry,jbarbuto/raven-python,beniwohli/apm-agent-python,vperron/sentry,someonehan/raven-python,korealerts1/sentry,arthurlogilab/raven-python,Kronuz/django-sentry,mvaled/sentry,beni55/sentry,llonchj/sentry,mitsuhiko/raven,beni55/sentry,tarkatronic/opbeat_python,BuildingLink/sentry,WoLpH/django-sentry,jbarbuto/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,nikolas/raven-python,gencer/sentry,recht/raven-python,alexm92/sentry,jbarbuto/raven-python,wong2/sentry,pauloschilling/sentry,lepture/raven-python,felixbuenemann/sentry,nicholasserra/sentry,icereval/raven-python,percipient/raven-python,jmagnusson/raven-python,Kronuz/django-sentry,Natim/sentry,icereval/raven-python,SilentCircle/sentry,songyi199111/sentry,argonemyth/sentry,beeftornado/sentry,camilonova/sentry,NickPresta/sentry,smarkets/raven-python,akalipetis/raven-python,recht/raven-python,chayapan/django-sentry,inspirehep/raven-python,patrys/opbeat_python,SilentCircle/sentry,zenefits/sentry,ronaldevers/raven-python,beniwohli/apm-agent-python,kevinlondon/sentry,jean/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,mvaled/sentry,BayanGroup/sentry,gencer/sentry,jmp0xf/raven-python,kevinastone/sentry,openlabs/raven,daevaorn/sentry,patrys/opbeat_python,wong2/sentry,JamesMura/sentry,WoLpH/django-sentry,Photonomie/raven-python,nikolas/raven-python,Kryz/sentry,looker/sentry,songyi199111/sentry,ifduyue/sentry,NickPresta/sentry,1tush/sentry,mitsuhiko/sentry,fotinakis/sentry,tarkatronic/opbeat_python,dirtycoder/opbeat_python,mvaled/sentry,vperron/sentry,korealerts1/sentry,dirtycoder/opbeat_python,jean/sentry,nikolas/raven-python,alex/sentry,jmagnusson/raven-python,percipient/raven-python,argonemyth/sentry,hongliang5623/sentry,JackDanger/sentry,kevinlondon/sentry,pauloschilling/sentry,fuziontech/sentry,tbarbugli/sentry_fork,rdio/sentry,smarkets/raven-python,looker/sentry,BuildingLink/sentry,camilonova/sentry,JTCunning/sentry,ifduyue/sentry,someonehan/raven-python,jean/sentry,Kryz/sentry,hzy/raven-python,arthurlogilab/raven-python,JamesMura/sentry,felixbuenemann/sentry,ronaldevers/raven-python,fuziontech/sentry,ifduyue/sentry,llonchj/sentry,SilentCircle/sentry,1tush/sentry,kevinastone/sentry,looker/sentry,zenefits/sentry,pauloschilling/sentry,alex/sentry,jbarbuto/raven-python,looker/sentry,collective/mr.poe,gg7/sentry,felixbuenemann/sentry,1tush/sentry,jean/sentry,ronaldevers/raven-python,Natim/sentry,BuildingLink/sentry,daikeren/opbeat_python,Natim/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,wujuguang/sentry,wujuguang/sentry,akalipetis/raven-python,JackDanger/sentry,BayanGroup/sentry,arthurlogilab/raven-python,boneyao/sentry,BayanGroup/sentry,mitsuhiko/raven,tbarbugli/sentry_fork,dcramer/sentry-old,Goldmund-Wyldebeast-Wunderliebe/raven-python,daikeren/opbeat_python,ticosax/opbeat_python,beeftornado/sentry,jmp0xf/raven-python,dbravender/raven-python,jokey2k/sentry,drcapulet/sentry,WoLpH/django-sentry,ticosax/opbeat_python,fotinakis/sentry,primepix/django-sentry,dcramer/sentry-old,BuildingLink/sentry,ewdurbin/raven-python,mvaled/sentry,ngonzalvez/sentry,llonchj/sentry,beni55/sentry,smarkets/raven-python,TedaLIEz/sentry,lepture/raven-python,hongliang5623/sentry,primepix/django-sentry,hzy/raven-python,akheron/raven-python,lopter/raven-python-old,Kronuz/django-sentry,nicholasserra/sentry,daevaorn/sentry,TedaLIEz/sentry,fotinakis/sentry,rdio/sentry,hzy/raven-python,vperron/sentry,imankulov/sentry,ifduyue/sentry,mvaled/sentry,korealerts1/sentry,ngonzalvez/sentry,tbarbugli/sentry_fork,rdio/sentry,getsentry/raven-python,chayapan/django-sentry,rdio/sentry,johansteffner/raven-python,getsentry/raven-python,jean/sentry,TedaLIEz/sentry,patrys/opbeat_python,ngonzalvez/sentry,camilonova/sentry,JackDanger/sentry,smarkets/raven-python,akheron/raven-python,kevinastone/sentry,Photonomie/raven-python,JamesMura/sentry,wujuguang/sentry,ticosax/opbeat_python,inspirehep/raven-python,beeftornado/sentry,argonemyth/sentry,beniwohli/apm-agent-python,tarkatronic/opbeat_python,ewdurbin/raven-python,fuziontech/sentry,zenefits/sentry,songyi199111/sentry,jokey2k/sentry,daikeren/opbeat_python,gencer/sentry,patrys/opbeat_python,dirtycoder/opbeat_python,icereval/raven-python,getsentry/raven-python,jmp0xf/raven-python,danriti/raven-python,nicholasserra/sentry,inspirehep/raven-python,wong2/sentry,gg7/sentry,imankulov/sentry,percipient/raven-python,johansteffner/raven-python,ifduyue/sentry,beniwohli/apm-agent-python,danriti/raven-python,alexm92/sentry,fotinakis/sentry,lepture/raven-python,akalipetis/raven-python,primepix/django-sentry,jmagnusson/raven-python,jokey2k/sentry,alexm92/sentry,JamesMura/sentry,drcapulet/sentry,kevinlondon/sentry,dbravender/raven-python,mitsuhiko/sentry | import logging
+ import sys
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
+ # Avoid typical config issues by overriding loggers behavior
+ if record.name == 'sentry.errors':
+ print >> sys.stderr, record.message
+ return
+
get_client().create_from_record(record) | Add a safety net for recursive logging | ## Code Before:
import logging
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
get_client().create_from_record(record)
## Instruction:
Add a safety net for recursive logging
## Code After:
import logging
import sys
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
# Avoid typical config issues by overriding loggers behavior
if record.name == 'sentry.errors':
print >> sys.stderr, record.message
return
get_client().create_from_record(record) | import logging
+ import sys
class SentryHandler(logging.Handler):
def emit(self, record):
from sentry.client.models import get_client
+ # Avoid typical config issues by overriding loggers behavior
+ if record.name == 'sentry.errors':
+ print >> sys.stderr, record.message
+ return
+
get_client().create_from_record(record) |
f4f439f24dceb0c68f05a90196b3e4b525d1aa7a | setup.py | setup.py |
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
)
|
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: DFSG approved',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Indexing/Search',
'Topic :: Software Development :: Libraries'],
)
| Add some trove classifiers to the package metadata | Add some trove classifiers to the package metadata
| Python | mit | rlskoeser/sunburnt,pixbuffer/sunburnt-spatial,anmar/sunburnt,rlskoeser/sunburnt,pixbuffer/sunburnt-spatial,qmssof/sunburnt,tow/sunburnt,anmar/sunburnt |
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
+ classifiers=[
+ 'Development Status :: 3 - Alpha',
+ 'Intended Audience :: Developers',
+ 'License :: DFSG approved',
+ 'Programming Language :: Python',
+ 'Topic :: Internet :: WWW/HTTP :: Indexing/Search',
+ 'Topic :: Software Development :: Libraries'],
)
| Add some trove classifiers to the package metadata | ## Code Before:
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
)
## Instruction:
Add some trove classifiers to the package metadata
## Code After:
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: DFSG approved',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Indexing/Search',
'Topic :: Software Development :: Libraries'],
)
|
import distutils.core
distutils.core.setup(
name='sunburnt',
version='0.4',
description='Python interface to Solr',
author='Toby White',
author_email='toby@timetric.com',
packages=['sunburnt'],
requires=['httplib2', 'lxml', 'pytz'],
license='WTFPL',
+ classifiers=[
+ 'Development Status :: 3 - Alpha',
+ 'Intended Audience :: Developers',
+ 'License :: DFSG approved',
+ 'Programming Language :: Python',
+ 'Topic :: Internet :: WWW/HTTP :: Indexing/Search',
+ 'Topic :: Software Development :: Libraries'],
) |
e4a3d3c273a1b7e26e9fdcf7f8da060bf127f27e | examples/django_project/django_project/tests.py | examples/django_project/django_project/tests.py | import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main()
| import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
# Note this tests should fail unless the snapshot-update command line
# option is specified. Run `python manage.py test --snapshot-update`.
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main()
| Add note for test_api_me django-project test | Add note for test_api_me django-project test
| Python | mit | syrusakbary/snapshottest | import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
+ # Note this tests should fail unless the snapshot-update command line
+ # option is specified. Run `python manage.py test --snapshot-update`.
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main()
| Add note for test_api_me django-project test | ## Code Before:
import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main()
## Instruction:
Add note for test_api_me django-project test
## Code After:
import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
# Note this tests should fail unless the snapshot-update command line
# option is specified. Run `python manage.py test --snapshot-update`.
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main()
| import unittest
from datetime import datetime
from snapshottest.django import TestCase
def api_client_get(url):
return {
'url': url,
}
class TestDemo(TestCase):
def test_api_me(self):
+ # Note this tests should fail unless the snapshot-update command line
+ # option is specified. Run `python manage.py test --snapshot-update`.
now = datetime.now().isoformat()
my_api_response = api_client_get('/' + now)
self.assertMatchSnapshot(my_api_response)
if __name__ == '__main__':
unittest.main() |
69fc2eccaa88189fd0de86d11206fa24d1508819 | tools/np_suppressions.py | tools/np_suppressions.py | suppressions = [
[ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
[ ".*/multiarray/calculation\.", "PyArray_Std" ],
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
[ ".*/multiarray/common\.", "PyCapsule_Check" ],
]
| suppressions = [
# This one cannot be covered by any Python language test because there is
# no code pathway to it. But it is part of the C API, so must not be
# excised from the code.
[ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
[ r".*/multiarray/calculation\.", "PyArray_Std" ],
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
[ r".*/multiarray/common\.", "PyCapsule_Check" ],
]
| Add documentation on one assertion, convert RE's to raw strings. | Add documentation on one assertion, convert RE's to raw strings.
| Python | bsd-3-clause | teoliphant/numpy-refactor,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,teoliphant/numpy-refactor | suppressions = [
+ # This one cannot be covered by any Python language test because there is
+ # no code pathway to it. But it is part of the C API, so must not be
+ # excised from the code.
- [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
+ [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
- [ ".*/multiarray/calculation\.", "PyArray_Std" ],
+ [ r".*/multiarray/calculation\.", "PyArray_Std" ],
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
- [ ".*/multiarray/common\.", "PyCapsule_Check" ],
+ [ r".*/multiarray/common\.", "PyCapsule_Check" ],
]
| Add documentation on one assertion, convert RE's to raw strings. | ## Code Before:
suppressions = [
[ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
[ ".*/multiarray/calculation\.", "PyArray_Std" ],
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
[ ".*/multiarray/common\.", "PyCapsule_Check" ],
]
## Instruction:
Add documentation on one assertion, convert RE's to raw strings.
## Code After:
suppressions = [
# This one cannot be covered by any Python language test because there is
# no code pathway to it. But it is part of the C API, so must not be
# excised from the code.
[ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
[ r".*/multiarray/calculation\.", "PyArray_Std" ],
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
[ r".*/multiarray/common\.", "PyCapsule_Check" ],
]
| suppressions = [
+ # This one cannot be covered by any Python language test because there is
+ # no code pathway to it. But it is part of the C API, so must not be
+ # excised from the code.
- [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
+ [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ],
? +
# PyArray_Std trivially forwards to and appears to be superceded by
# __New_PyArray_Std, which is exercised by the test framework.
- [ ".*/multiarray/calculation\.", "PyArray_Std" ],
+ [ r".*/multiarray/calculation\.", "PyArray_Std" ],
? +
# PyCapsule_Check is declared in a header, and used in
# multiarray/ctors.c. So it isn't really untested.
- [ ".*/multiarray/common\.", "PyCapsule_Check" ],
+ [ r".*/multiarray/common\.", "PyCapsule_Check" ],
? +
] |
9f3abe5077fce0a2d7323a769fc063fca5b7aca8 | tests/test_bawlerd.py | tests/test_bawlerd.py | import os
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
| import io
import os
from textwrap import dedent
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
def test__load_file(self):
config = bawlerd.conf._load_file(io.StringIO(dedent("""\
logging:
formatters:
standard:
format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\"
handlers:
default:
level: "INFO"
formatter: standard
class: logging.StreamHandler
loggers:
"":
handlers: ["default"]
level: INFO
propagate: True
""")))
assert 'logging' in config
| Add simple test for _load_file | Add simple test for _load_file
Signed-off-by: Michal Kuffa <005ee1c97edba97d164343c993afee612ac25a0c@gmail.com>
| Python | bsd-3-clause | beezz/pg_bawler,beezz/pg_bawler | + import io
import os
+ from textwrap import dedent
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
+ def test__load_file(self):
+ config = bawlerd.conf._load_file(io.StringIO(dedent("""\
+ logging:
+ formatters:
+ standard:
+ format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\"
+ handlers:
+ default:
+ level: "INFO"
+ formatter: standard
+ class: logging.StreamHandler
+ loggers:
+ "":
+ handlers: ["default"]
+ level: INFO
+ propagate: True
+ """)))
+ assert 'logging' in config
+ | Add simple test for _load_file | ## Code Before:
import os
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
## Instruction:
Add simple test for _load_file
## Code After:
import io
import os
from textwrap import dedent
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
def test__load_file(self):
config = bawlerd.conf._load_file(io.StringIO(dedent("""\
logging:
formatters:
standard:
format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\"
handlers:
default:
level: "INFO"
formatter: standard
class: logging.StreamHandler
loggers:
"":
handlers: ["default"]
level: INFO
propagate: True
""")))
assert 'logging' in config
| + import io
import os
+ from textwrap import dedent
from pg_bawler import bawlerd
class TestBawlerdConfig:
def test_build_config_location_list(self):
assert not bawlerd.conf.build_config_location_list(locations=())
user_conf = os.path.join(
os.path.expanduser('~'),
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
system_conf = os.path.join(
'/etc/pg_bawler',
bawlerd.conf.DEFAULT_CONFIG_FILENAME)
assert user_conf in bawlerd.conf.build_config_location_list()
assert system_conf in bawlerd.conf.build_config_location_list()
+
+ def test__load_file(self):
+ config = bawlerd.conf._load_file(io.StringIO(dedent("""\
+ logging:
+ formatters:
+ standard:
+ format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\"
+ handlers:
+ default:
+ level: "INFO"
+ formatter: standard
+ class: logging.StreamHandler
+ loggers:
+ "":
+ handlers: ["default"]
+ level: INFO
+ propagate: True
+ """)))
+ assert 'logging' in config |
1709c602b8a423d1eee6521c5e74987db0fc8b81 | fancypages/contrib/oscar_fancypages/mixins.py | fancypages/contrib/oscar_fancypages/mixins.py | from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
context_object_name = 'fancypage'
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
ctx[self.context_object_name] = getattr(self, self.page_attr_name)
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx
| from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
context_object_name = 'products'
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
ctx['fancypage'] = getattr(self, self.page_attr_name)
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx
| Change context object for product list view in Oscar contrib | Change context object for product list view in Oscar contrib
| Python | bsd-3-clause | tangentlabs/django-fancypages,tangentlabs/django-fancypages,tangentlabs/django-fancypages,socradev/django-fancypages,socradev/django-fancypages,socradev/django-fancypages | from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
- context_object_name = 'fancypage'
+ context_object_name = 'products'
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
- ctx[self.context_object_name] = getattr(self, self.page_attr_name)
+ ctx['fancypage'] = getattr(self, self.page_attr_name)
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx
| Change context object for product list view in Oscar contrib | ## Code Before:
from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
context_object_name = 'fancypage'
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
ctx[self.context_object_name] = getattr(self, self.page_attr_name)
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx
## Instruction:
Change context object for product list view in Oscar contrib
## Code After:
from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
context_object_name = 'products'
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
ctx['fancypage'] = getattr(self, self.page_attr_name)
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx
| from ... import mixins
class OscarFancyPageMixin(mixins.FancyPageMixin):
node_attr_name = 'category'
slug_url_kwarg = 'category_slug'
- context_object_name = 'fancypage'
? ^^^ ^^^^^
+ context_object_name = 'products'
? ^^^^^ ^^
def get_context_data(self, **kwargs):
ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs)
- ctx[self.context_object_name] = getattr(self, self.page_attr_name)
? ^ ^^^^^^^^^^^^^^^^^^^^^^
+ ctx['fancypage'] = getattr(self, self.page_attr_name)
? ^^^^^^^^^ ^
if self.category:
ctx['object'] = self.category
for container in self.category.page.containers.all():
ctx[container.name] = container
return ctx |
0dabc858976197459cfe71fe1a4a8a85c181db75 | django_localflavor_ie/ie_counties.py | django_localflavor_ie/ie_counties.py | from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
('antrim', _('Antrim')),
('armagh', _('Armagh')),
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
('derry', _('Derry')),
('donegal', _('Donegal')),
('down', _('Down')),
('dublin', _('Dublin')),
('fermanagh', _('Fermanagh')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
('tyrone', _('Tyrone')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
)
| from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
('donegal', _('Donegal')),
('dublin', _('Dublin')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
)
| Remove Northern Irish counties. These are part of the UK, not Ireland | Remove Northern Irish counties. These are part of the UK, not Ireland
| Python | bsd-3-clause | martinogden/django-localflavor-ie | from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
- ('antrim', _('Antrim')),
- ('armagh', _('Armagh')),
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
- ('derry', _('Derry')),
('donegal', _('Donegal')),
- ('down', _('Down')),
('dublin', _('Dublin')),
- ('fermanagh', _('Fermanagh')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
- ('tyrone', _('Tyrone')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
)
| Remove Northern Irish counties. These are part of the UK, not Ireland | ## Code Before:
from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
('antrim', _('Antrim')),
('armagh', _('Armagh')),
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
('derry', _('Derry')),
('donegal', _('Donegal')),
('down', _('Down')),
('dublin', _('Dublin')),
('fermanagh', _('Fermanagh')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
('tyrone', _('Tyrone')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
)
## Instruction:
Remove Northern Irish counties. These are part of the UK, not Ireland
## Code After:
from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
('donegal', _('Donegal')),
('dublin', _('Dublin')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
)
| from django.utils.translation import ugettext_lazy as _
IE_COUNTY_CHOICES = (
- ('antrim', _('Antrim')),
- ('armagh', _('Armagh')),
('carlow', _('Carlow')),
('cavan', _('Cavan')),
('clare', _('Clare')),
('cork', _('Cork')),
- ('derry', _('Derry')),
('donegal', _('Donegal')),
- ('down', _('Down')),
('dublin', _('Dublin')),
- ('fermanagh', _('Fermanagh')),
('galway', _('Galway')),
('kerry', _('Kerry')),
('kildare', _('Kildare')),
('kilkenny', _('Kilkenny')),
('laois', _('Laois')),
('leitrim', _('Leitrim')),
('limerick', _('Limerick')),
('longford', _('Longford')),
('louth', _('Louth')),
('mayo', _('Mayo')),
('meath', _('Meath')),
('monaghan', _('Monaghan')),
('offaly', _('Offaly')),
('roscommon', _('Roscommon')),
('sligo', _('Sligo')),
('tipperary', _('Tipperary')),
- ('tyrone', _('Tyrone')),
('waterford', _('Waterford')),
('westmeath', _('Westmeath')),
('wexford', _('Wexford')),
('wicklow', _('Wicklow')),
) |
dab7eaadbc6fc0dd867358b096a846ec39bc0440 | pnnl/models/__init__.py | pnnl/models/__init__.py | import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
| import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
return q
| Add return statement to Model.get_q | Add return statement to Model.get_q
| Python | bsd-3-clause | VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications | import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
+ return q
| Add return statement to Model.get_q | ## Code Before:
import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
## Instruction:
Add return statement to Model.get_q
## Code After:
import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
return q
| import importlib
import logging
from volttron.platform.agent import utils
_log = logging.getLogger(__name__)
utils.setup_logging()
__version__ = "0.1"
__all__ = ['Model']
class Model(object):
def __init__(self, config, **kwargs):
base_module = "volttron.pnnl.models."
try:
model_type = config["model_type"]
except KeyError as e:
_log.exception("Missing Model Type key: {}".format(e))
raise e
_file, model_type = model_type.split(".")
module = importlib.import_module(base_module + _file)
model_class = getattr(module, model_type)
self.model = model_class(config, self)
def get_q(self, _set, sched_index, market_index, occupied):
q = self.model.predict(_set, sched_index, market_index, occupied)
+ return q |
e0d631b4aab431c31689ccd7aa6ac92d95e32e80 | tests/test_frontend.py | tests/test_frontend.py | import os
from tvrenamr.cli import helpers
from .base import BaseTest
class TestFrontEnd(BaseTest):
def setup(self):
super(TestFrontEnd, self).setup()
self.config = helpers.get_config()
def test_passing_current_dir_makes_file_list_a_list(self):
assert isinstance(helpers.build_file_list([self.files]), list)
def test_setting_recursive_adds_all_files_below_the_folder(self):
new_folders = ('herp', 'derp', 'test')
os.makedirs(os.path.join(self.files, *new_folders))
def build_folder(folder):
new_files = ('foo', 'bar', 'blah')
for fn in new_files:
with open(os.path.join(self.files, folder, fn), 'w') as f:
f.write('')
build_folder('herp')
build_folder('herp/derp')
build_folder('herp/derp/test')
file_list = helpers.build_file_list([self.files], recursive=True)
for root, dirs, files in os.walk(self.files):
for fn in files:
assert os.path.join(root, fn) in file_list
def test_ignoring_files(self):
ignore = self.random_files(self.files)
file_list = helpers.build_file_list([self.files], ignore_filelist=ignore)
assert all(fn not in file_list for fn in ignore)
| import collections
import os
import sys
from tvrenamr.cli import helpers
from .utils import random_files
def test_passing_current_dir_makes_file_list_a_list(files):
file_list = helpers.build_file_list([files])
assert isinstance(file_list, collections.Iterable)
PY3 = sys.version_info[0] == 3
string_type = str if PY3 else basestring
text_type = str if PY3 else unicode
assert not isinstance(file_list, (string_type, text_type))
def test_setting_recursive_adds_all_files_below_the_folder(files):
new_folders = ('herp', 'derp', 'test')
os.makedirs(os.path.join(files, *new_folders))
def build_folder(folder):
new_files = ('foo', 'bar', 'blah')
for fn in new_files:
with open(os.path.join(files, folder, fn), 'w') as f:
f.write('')
build_folder('herp')
build_folder('herp/derp')
build_folder('herp/derp/test')
file_list = helpers.build_file_list([files], recursive=True)
for root, dirs, files in os.walk(files):
for fn in files:
assert (root, fn) in file_list
def test_ignoring_files(files):
ignore = random_files(files)
file_list = helpers.build_file_list([files], ignore_filelist=ignore)
assert all(fn not in file_list for fn in ignore)
| Move to function only tests & fix test for generator based build_file_list | Move to function only tests & fix test for generator based build_file_list
build_file_list is a generator now so we need to make sure it returns an
iterable but not a string.
| Python | mit | wintersandroid/tvrenamr,ghickman/tvrenamr | + import collections
import os
+ import sys
from tvrenamr.cli import helpers
- from .base import BaseTest
+ from .utils import random_files
+ def test_passing_current_dir_makes_file_list_a_list(files):
+ file_list = helpers.build_file_list([files])
- class TestFrontEnd(BaseTest):
- def setup(self):
- super(TestFrontEnd, self).setup()
- self.config = helpers.get_config()
+ assert isinstance(file_list, collections.Iterable)
- def test_passing_current_dir_makes_file_list_a_list(self):
- assert isinstance(helpers.build_file_list([self.files]), list)
- def test_setting_recursive_adds_all_files_below_the_folder(self):
- new_folders = ('herp', 'derp', 'test')
- os.makedirs(os.path.join(self.files, *new_folders))
+ PY3 = sys.version_info[0] == 3
+ string_type = str if PY3 else basestring
+ text_type = str if PY3 else unicode
+ assert not isinstance(file_list, (string_type, text_type))
- def build_folder(folder):
- new_files = ('foo', 'bar', 'blah')
- for fn in new_files:
- with open(os.path.join(self.files, folder, fn), 'w') as f:
- f.write('')
- build_folder('herp')
- build_folder('herp/derp')
- build_folder('herp/derp/test')
- file_list = helpers.build_file_list([self.files], recursive=True)
- for root, dirs, files in os.walk(self.files):
- for fn in files:
- assert os.path.join(root, fn) in file_list
+ def test_setting_recursive_adds_all_files_below_the_folder(files):
+ new_folders = ('herp', 'derp', 'test')
+ os.makedirs(os.path.join(files, *new_folders))
- def test_ignoring_files(self):
- ignore = self.random_files(self.files)
- file_list = helpers.build_file_list([self.files], ignore_filelist=ignore)
- assert all(fn not in file_list for fn in ignore)
+ def build_folder(folder):
+ new_files = ('foo', 'bar', 'blah')
+ for fn in new_files:
+ with open(os.path.join(files, folder, fn), 'w') as f:
+ f.write('')
+ build_folder('herp')
+ build_folder('herp/derp')
+ build_folder('herp/derp/test')
+ file_list = helpers.build_file_list([files], recursive=True)
+ for root, dirs, files in os.walk(files):
+ for fn in files:
+ assert (root, fn) in file_list
+
+
+ def test_ignoring_files(files):
+ ignore = random_files(files)
+ file_list = helpers.build_file_list([files], ignore_filelist=ignore)
+ assert all(fn not in file_list for fn in ignore)
+ | Move to function only tests & fix test for generator based build_file_list | ## Code Before:
import os
from tvrenamr.cli import helpers
from .base import BaseTest
class TestFrontEnd(BaseTest):
def setup(self):
super(TestFrontEnd, self).setup()
self.config = helpers.get_config()
def test_passing_current_dir_makes_file_list_a_list(self):
assert isinstance(helpers.build_file_list([self.files]), list)
def test_setting_recursive_adds_all_files_below_the_folder(self):
new_folders = ('herp', 'derp', 'test')
os.makedirs(os.path.join(self.files, *new_folders))
def build_folder(folder):
new_files = ('foo', 'bar', 'blah')
for fn in new_files:
with open(os.path.join(self.files, folder, fn), 'w') as f:
f.write('')
build_folder('herp')
build_folder('herp/derp')
build_folder('herp/derp/test')
file_list = helpers.build_file_list([self.files], recursive=True)
for root, dirs, files in os.walk(self.files):
for fn in files:
assert os.path.join(root, fn) in file_list
def test_ignoring_files(self):
ignore = self.random_files(self.files)
file_list = helpers.build_file_list([self.files], ignore_filelist=ignore)
assert all(fn not in file_list for fn in ignore)
## Instruction:
Move to function only tests & fix test for generator based build_file_list
## Code After:
import collections
import os
import sys
from tvrenamr.cli import helpers
from .utils import random_files
def test_passing_current_dir_makes_file_list_a_list(files):
file_list = helpers.build_file_list([files])
assert isinstance(file_list, collections.Iterable)
PY3 = sys.version_info[0] == 3
string_type = str if PY3 else basestring
text_type = str if PY3 else unicode
assert not isinstance(file_list, (string_type, text_type))
def test_setting_recursive_adds_all_files_below_the_folder(files):
new_folders = ('herp', 'derp', 'test')
os.makedirs(os.path.join(files, *new_folders))
def build_folder(folder):
new_files = ('foo', 'bar', 'blah')
for fn in new_files:
with open(os.path.join(files, folder, fn), 'w') as f:
f.write('')
build_folder('herp')
build_folder('herp/derp')
build_folder('herp/derp/test')
file_list = helpers.build_file_list([files], recursive=True)
for root, dirs, files in os.walk(files):
for fn in files:
assert (root, fn) in file_list
def test_ignoring_files(files):
ignore = random_files(files)
file_list = helpers.build_file_list([files], ignore_filelist=ignore)
assert all(fn not in file_list for fn in ignore)
| + import collections
import os
+ import sys
from tvrenamr.cli import helpers
- from .base import BaseTest
+ from .utils import random_files
+ def test_passing_current_dir_makes_file_list_a_list(files):
+ file_list = helpers.build_file_list([files])
- class TestFrontEnd(BaseTest):
- def setup(self):
- super(TestFrontEnd, self).setup()
- self.config = helpers.get_config()
+ assert isinstance(file_list, collections.Iterable)
- def test_passing_current_dir_makes_file_list_a_list(self):
- assert isinstance(helpers.build_file_list([self.files]), list)
- def test_setting_recursive_adds_all_files_below_the_folder(self):
- new_folders = ('herp', 'derp', 'test')
- os.makedirs(os.path.join(self.files, *new_folders))
+ PY3 = sys.version_info[0] == 3
+ string_type = str if PY3 else basestring
+ text_type = str if PY3 else unicode
+ assert not isinstance(file_list, (string_type, text_type))
- def build_folder(folder):
- new_files = ('foo', 'bar', 'blah')
- for fn in new_files:
- with open(os.path.join(self.files, folder, fn), 'w') as f:
- f.write('')
- build_folder('herp')
- build_folder('herp/derp')
- build_folder('herp/derp/test')
- file_list = helpers.build_file_list([self.files], recursive=True)
- for root, dirs, files in os.walk(self.files):
- for fn in files:
- assert os.path.join(root, fn) in file_list
+ def test_setting_recursive_adds_all_files_below_the_folder(files):
+ new_folders = ('herp', 'derp', 'test')
+ os.makedirs(os.path.join(files, *new_folders))
+
+ def build_folder(folder):
+ new_files = ('foo', 'bar', 'blah')
+ for fn in new_files:
+ with open(os.path.join(files, folder, fn), 'w') as f:
+ f.write('')
+ build_folder('herp')
+ build_folder('herp/derp')
+ build_folder('herp/derp/test')
+ file_list = helpers.build_file_list([files], recursive=True)
+ for root, dirs, files in os.walk(files):
+ for fn in files:
+ assert (root, fn) in file_list
+
+
- def test_ignoring_files(self):
? ---- ---
+ def test_ignoring_files(files):
? ++++
- ignore = self.random_files(self.files)
? ---- ----- -----
+ ignore = random_files(files)
- file_list = helpers.build_file_list([self.files], ignore_filelist=ignore)
? ---- -----
+ file_list = helpers.build_file_list([files], ignore_filelist=ignore)
- assert all(fn not in file_list for fn in ignore)
? ----
+ assert all(fn not in file_list for fn in ignore) |
e2d51e23f530202b82ba13ae11c686deb1388435 | prototype/BioID.py | prototype/BioID.py |
import re
import json
import mmap
class BioID:
defs = None
def __init__(self, defpath):
with open(defpath, "r") as deffile:
conts = deffile.read()
self.defs = json.loads(conts)["formats"]
@classmethod
def identify(cls, files):
recog = {}
for file in files:
with open(file, "r") as infile:
buff = infile.read()
mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
if len(buff) == 0:
recog[file] = "empty" # Empty files have no format :)
continue
for fdef in cls.defs:
matched = True
if "regexen" in fdef:
for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
matched = False
break
if "bytes" in fdef:
for bytes in fdef["bytes"]:
if mem_map.find(bytes.decode("string_escape")) == -1:
matched = False
break
if matched:
recog[file] = fdef["name"]
break
mem_map.close()
if file not in recog:
recog[file] = "unrecognized"
return recog |
import re
import json
import mmap
class BioID:
defs = None
def __init__(self, defpath):
with open(defpath, "r") as deffile:
conts = deffile.read()
self.defs = json.loads(conts)["formats"]
@classmethod
def identify(cls, files):
recog = {}
for file in files:
with open(file, "r") as infile:
buff = infile.read()
mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
if len(buff) == 0:
recog[file] = "empty" # Empty files have no format :)
continue
for fdef in cls.defs:
matched = True
if "regexen" in fdef:
for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
matched = False
break
if "bytes" in fdef:
for bytes in fdef["bytes"]:
if mem_map.find(bytes.decode("string_escape")) == -1:
matched = False
break
if matched:
recog[file] = fdef["name"]
break
mem_map.close()
if file not in recog:
recog[file] = "unrecognized"
return recog | Indent return in identify class. | Indent return in identify class.
| Python | mit | LeeBergstrand/BioMagick,LeeBergstrand/BioMagick |
import re
import json
import mmap
class BioID:
- defs = None
+ defs = None
- def __init__(self, defpath):
+ def __init__(self, defpath):
- with open(defpath, "r") as deffile:
+ with open(defpath, "r") as deffile:
- conts = deffile.read()
+ conts = deffile.read()
- self.defs = json.loads(conts)["formats"]
+ self.defs = json.loads(conts)["formats"]
- @classmethod
+ @classmethod
- def identify(cls, files):
+ def identify(cls, files):
- recog = {}
+ recog = {}
- for file in files:
+ for file in files:
- with open(file, "r") as infile:
+ with open(file, "r") as infile:
- buff = infile.read()
+ buff = infile.read()
- mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
+ mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
- if len(buff) == 0:
+ if len(buff) == 0:
- recog[file] = "empty" # Empty files have no format :)
+ recog[file] = "empty" # Empty files have no format :)
- continue
+ continue
- for fdef in cls.defs:
- matched = True
- if "regexen" in fdef:
- for regex in fdef["regexen"]:
+ for fdef in cls.defs:
+ matched = True
+ if "regexen" in fdef:
+ for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
- matched = False
- break
- if "bytes" in fdef:
- for bytes in fdef["bytes"]:
+ matched = False
+ break
+ if "bytes" in fdef:
+ for bytes in fdef["bytes"]:
- if mem_map.find(bytes.decode("string_escape")) == -1:
+ if mem_map.find(bytes.decode("string_escape")) == -1:
- matched = False
- break
- if matched:
- recog[file] = fdef["name"]
- break
+ matched = False
+ break
+ if matched:
+ recog[file] = fdef["name"]
+ break
- mem_map.close()
+ mem_map.close()
- if file not in recog:
+ if file not in recog:
- recog[file] = "unrecognized"
+ recog[file] = "unrecognized"
- return recog
+ return recog | Indent return in identify class. | ## Code Before:
import re
import json
import mmap
class BioID:
defs = None
def __init__(self, defpath):
with open(defpath, "r") as deffile:
conts = deffile.read()
self.defs = json.loads(conts)["formats"]
@classmethod
def identify(cls, files):
recog = {}
for file in files:
with open(file, "r") as infile:
buff = infile.read()
mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
if len(buff) == 0:
recog[file] = "empty" # Empty files have no format :)
continue
for fdef in cls.defs:
matched = True
if "regexen" in fdef:
for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
matched = False
break
if "bytes" in fdef:
for bytes in fdef["bytes"]:
if mem_map.find(bytes.decode("string_escape")) == -1:
matched = False
break
if matched:
recog[file] = fdef["name"]
break
mem_map.close()
if file not in recog:
recog[file] = "unrecognized"
return recog
## Instruction:
Indent return in identify class.
## Code After:
import re
import json
import mmap
class BioID:
defs = None
def __init__(self, defpath):
with open(defpath, "r") as deffile:
conts = deffile.read()
self.defs = json.loads(conts)["formats"]
@classmethod
def identify(cls, files):
recog = {}
for file in files:
with open(file, "r") as infile:
buff = infile.read()
mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
if len(buff) == 0:
recog[file] = "empty" # Empty files have no format :)
continue
for fdef in cls.defs:
matched = True
if "regexen" in fdef:
for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
matched = False
break
if "bytes" in fdef:
for bytes in fdef["bytes"]:
if mem_map.find(bytes.decode("string_escape")) == -1:
matched = False
break
if matched:
recog[file] = fdef["name"]
break
mem_map.close()
if file not in recog:
recog[file] = "unrecognized"
return recog |
import re
import json
import mmap
class BioID:
- defs = None
? ^^^^
+ defs = None
? ^
- def __init__(self, defpath):
? ^^^^
+ def __init__(self, defpath):
? ^
- with open(defpath, "r") as deffile:
? ^^^^^^^^
+ with open(defpath, "r") as deffile:
? ^^
- conts = deffile.read()
+ conts = deffile.read()
- self.defs = json.loads(conts)["formats"]
? ^^^^^^^^
+ self.defs = json.loads(conts)["formats"]
? ^^
- @classmethod
? ^^^^
+ @classmethod
? ^
- def identify(cls, files):
? ^^^^
+ def identify(cls, files):
? ^
- recog = {}
+ recog = {}
- for file in files:
? ^^^^^^^^
+ for file in files:
? ^^
- with open(file, "r") as infile:
? ^^^^^^^^^^^^
+ with open(file, "r") as infile:
? ^^^
- buff = infile.read()
+ buff = infile.read()
- mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
? ^^^^^^^^^^^^^^^^
+ mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ)
? ^^^^
- if len(buff) == 0:
+ if len(buff) == 0:
- recog[file] = "empty" # Empty files have no format :)
? ^^^^^^^^^^^^^^^^
+ recog[file] = "empty" # Empty files have no format :)
? ^^^^
- continue
+ continue
- for fdef in cls.defs:
- matched = True
- if "regexen" in fdef:
- for regex in fdef["regexen"]:
+ for fdef in cls.defs:
+ matched = True
+ if "regexen" in fdef:
+ for regex in fdef["regexen"]:
if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE):
- matched = False
- break
- if "bytes" in fdef:
- for bytes in fdef["bytes"]:
+ matched = False
+ break
+ if "bytes" in fdef:
+ for bytes in fdef["bytes"]:
- if mem_map.find(bytes.decode("string_escape")) == -1:
? ^^^^^^^^^^^^^^^^^^^^^^^^
+ if mem_map.find(bytes.decode("string_escape")) == -1:
? ^^^^^
- matched = False
- break
- if matched:
- recog[file] = fdef["name"]
- break
+ matched = False
+ break
+ if matched:
+ recog[file] = fdef["name"]
+ break
- mem_map.close()
+ mem_map.close()
- if file not in recog:
? ^^^^^^^^^^^^
+ if file not in recog:
? ^^
- recog[file] = "unrecognized"
+ recog[file] = "unrecognized"
- return recog
+ return recog |
c814fe264c93dfa09276474960aa83cdb26e7754 | polyaxon/api/searches/serializers.py | polyaxon/api/searches/serializers.py | from rest_framework import serializers
from db.models.searches import Search
class SearchSerializer(serializers.ModelSerializer):
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
| from rest_framework import serializers
from rest_framework.exceptions import ValidationError
from api.utils.serializers.names import NamesMixin
from db.models.searches import Search
class SearchSerializer(serializers.ModelSerializer, NamesMixin):
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
def create(self, validated_data):
validated_data = self.validated_name(validated_data,
project=validated_data['project'],
query=Search.all)
try:
return super().create(validated_data)
except Exception as e:
raise ValidationError(e)
| Add graceful handling for creating search with similar names | Add graceful handling for creating search with similar names
| Python | apache-2.0 | polyaxon/polyaxon,polyaxon/polyaxon,polyaxon/polyaxon | from rest_framework import serializers
+ from rest_framework.exceptions import ValidationError
+ from api.utils.serializers.names import NamesMixin
from db.models.searches import Search
- class SearchSerializer(serializers.ModelSerializer):
+ class SearchSerializer(serializers.ModelSerializer, NamesMixin):
-
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
+ def create(self, validated_data):
+ validated_data = self.validated_name(validated_data,
+ project=validated_data['project'],
+ query=Search.all)
+ try:
+ return super().create(validated_data)
+ except Exception as e:
+ raise ValidationError(e)
+ | Add graceful handling for creating search with similar names | ## Code Before:
from rest_framework import serializers
from db.models.searches import Search
class SearchSerializer(serializers.ModelSerializer):
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
## Instruction:
Add graceful handling for creating search with similar names
## Code After:
from rest_framework import serializers
from rest_framework.exceptions import ValidationError
from api.utils.serializers.names import NamesMixin
from db.models.searches import Search
class SearchSerializer(serializers.ModelSerializer, NamesMixin):
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
def create(self, validated_data):
validated_data = self.validated_name(validated_data,
project=validated_data['project'],
query=Search.all)
try:
return super().create(validated_data)
except Exception as e:
raise ValidationError(e)
| from rest_framework import serializers
+ from rest_framework.exceptions import ValidationError
+ from api.utils.serializers.names import NamesMixin
from db.models.searches import Search
- class SearchSerializer(serializers.ModelSerializer):
+ class SearchSerializer(serializers.ModelSerializer, NamesMixin):
? ++++++++++++
-
class Meta:
model = Search
fields = ['id', 'name', 'query', 'meta']
+
+ def create(self, validated_data):
+ validated_data = self.validated_name(validated_data,
+ project=validated_data['project'],
+ query=Search.all)
+ try:
+ return super().create(validated_data)
+ except Exception as e:
+ raise ValidationError(e) |
0a2fa84285a586282d79146f85d9efba12a528dd | Parallel/Testing/Cxx/TestSockets.py | Parallel/Testing/Cxx/TestSockets.py | import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
sys.argv[4] ))
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
| import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
sys.argv[4] ))
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
sys.exit(os.WEXITSTATUS(retVal))
| Return code from script must reflect that of the test. | BUG: Return code from script must reflect that of the test.
| Python | bsd-3-clause | mspark93/VTK,jeffbaumes/jeffbaumes-vtk,demarle/VTK,sumedhasingla/VTK,sankhesh/VTK,mspark93/VTK,keithroe/vtkoptix,daviddoria/PointGraphsPhase1,SimVascular/VTK,collects/VTK,SimVascular/VTK,sumedhasingla/VTK,mspark93/VTK,biddisco/VTK,sgh/vtk,jmerkow/VTK,aashish24/VTK-old,demarle/VTK,demarle/VTK,aashish24/VTK-old,johnkit/vtk-dev,johnkit/vtk-dev,johnkit/vtk-dev,msmolens/VTK,hendradarwin/VTK,SimVascular/VTK,johnkit/vtk-dev,hendradarwin/VTK,cjh1/VTK,demarle/VTK,ashray/VTK-EVM,daviddoria/PointGraphsPhase1,sankhesh/VTK,keithroe/vtkoptix,Wuteyan/VTK,mspark93/VTK,spthaolt/VTK,candy7393/VTK,aashish24/VTK-old,jmerkow/VTK,sgh/vtk,gram526/VTK,Wuteyan/VTK,daviddoria/PointGraphsPhase1,sgh/vtk,spthaolt/VTK,keithroe/vtkoptix,sumedhasingla/VTK,hendradarwin/VTK,berendkleinhaneveld/VTK,ashray/VTK-EVM,berendkleinhaneveld/VTK,jeffbaumes/jeffbaumes-vtk,jmerkow/VTK,sankhesh/VTK,SimVascular/VTK,ashray/VTK-EVM,keithroe/vtkoptix,ashray/VTK-EVM,arnaudgelas/VTK,biddisco/VTK,candy7393/VTK,collects/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,jeffbaumes/jeffbaumes-vtk,naucoin/VTKSlicerWidgets,aashish24/VTK-old,Wuteyan/VTK,msmolens/VTK,jmerkow/VTK,cjh1/VTK,jeffbaumes/jeffbaumes-vtk,naucoin/VTKSlicerWidgets,jmerkow/VTK,johnkit/vtk-dev,msmolens/VTK,jmerkow/VTK,spthaolt/VTK,cjh1/VTK,spthaolt/VTK,msmolens/VTK,biddisco/VTK,cjh1/VTK,collects/VTK,arnaudgelas/VTK,Wuteyan/VTK,aashish24/VTK-old,spthaolt/VTK,hendradarwin/VTK,mspark93/VTK,naucoin/VTKSlicerWidgets,naucoin/VTKSlicerWidgets,jeffbaumes/jeffbaumes-vtk,SimVascular/VTK,Wuteyan/VTK,sankhesh/VTK,sankhesh/VTK,hendradarwin/VTK,arnaudgelas/VTK,daviddoria/PointGraphsPhase1,collects/VTK,spthaolt/VTK,Wuteyan/VTK,johnkit/vtk-dev,Wuteyan/VTK,gram526/VTK,candy7393/VTK,candy7393/VTK,candy7393/VTK,keithroe/vtkoptix,keithroe/vtkoptix,sumedhasingla/VTK,sankhesh/VTK,berendkleinhaneveld/VTK,candy7393/VTK,msmolens/VTK,demarle/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,msmolens/VTK,ashray/VTK-EVM,daviddoria/PointGraphsPhase1,msmolens/VTK,keithroe/vtkoptix,gram526/VTK,sgh/vtk,sgh/vtk,keithroe/vtkoptix,berendkleinhaneveld/VTK,demarle/VTK,sankhesh/VTK,candy7393/VTK,daviddoria/PointGraphsPhase1,collects/VTK,cjh1/VTK,jeffbaumes/jeffbaumes-vtk,collects/VTK,johnkit/vtk-dev,gram526/VTK,sumedhasingla/VTK,biddisco/VTK,ashray/VTK-EVM,mspark93/VTK,mspark93/VTK,sgh/vtk,demarle/VTK,arnaudgelas/VTK,hendradarwin/VTK,jmerkow/VTK,gram526/VTK,candy7393/VTK,SimVascular/VTK,jmerkow/VTK,aashish24/VTK-old,gram526/VTK,biddisco/VTK,mspark93/VTK,sankhesh/VTK,hendradarwin/VTK,ashray/VTK-EVM,biddisco/VTK,SimVascular/VTK,msmolens/VTK,naucoin/VTKSlicerWidgets,biddisco/VTK,gram526/VTK,cjh1/VTK,ashray/VTK-EVM,SimVascular/VTK,naucoin/VTKSlicerWidgets,arnaudgelas/VTK,arnaudgelas/VTK,spthaolt/VTK,gram526/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,demarle/VTK | import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
- os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
+ retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
- sys.argv[4] ))
+ sys.argv[4] ))
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
+ sys.exit(os.WEXITSTATUS(retVal))
- | Return code from script must reflect that of the test. | ## Code Before:
import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
sys.argv[4] ))
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
## Instruction:
Return code from script must reflect that of the test.
## Code After:
import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
sys.argv[4] ))
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
sys.exit(os.WEXITSTATUS(retVal))
| import os, sys, time
# Fork, run server in child, client in parent
pid = os.fork()
if pid == 0:
# exec the parent
os.execv(sys.argv[1], ('-D', sys.argv[3]))
else:
# wait a little to make sure that the server is ready
time.sleep(10)
# run the client
- os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
+ retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3],
? +++++++++
- sys.argv[4] ))
+ sys.argv[4] ))
? +++++++++
# in case the client fails, we need to kill the server
# or it will stay around
time.sleep(20)
os.kill(pid, 15)
-
+ sys.exit(os.WEXITSTATUS(retVal)) |
2f860583a99b88324b19b1118b4aea29a28ae90d | polling_stations/apps/data_collection/management/commands/import_portsmouth.py | polling_stations/apps/data_collection/management/commands/import_portsmouth.py | from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
elections = ["local.2018-05-03"]
csv_delimiter = "\t"
def address_record_to_dict(self, record):
if record.addressline6 == "PO1 5BZ":
return None
return super().address_record_to_dict(record)
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
if rec["internal_council_id"] == "3270":
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec
| from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
addresses_name = (
"local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
)
stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
elections = ["local.2019-05-02"]
csv_delimiter = "\t"
def address_record_to_dict(self, record):
rec = super().address_record_to_dict(record)
if record.addressline6 == "PO4 099":
rec["postcode"] = "PO4 0PL"
if record.property_urn.strip().lstrip("0") in [
"1775122942",
"1775122943",
"1775122944",
]:
rec["postcode"] = "PO5 2BZ"
return rec
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
if rec["internal_council_id"] == "3596":
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec
| Add import script for Portsmouth | Add import script for Portsmouth
Closes #1502
| Python | bsd-3-clause | DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations | from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
+ addresses_name = (
- addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
+ "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
+ )
- stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
+ stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
- elections = ["local.2018-05-03"]
+ elections = ["local.2019-05-02"]
csv_delimiter = "\t"
def address_record_to_dict(self, record):
+ rec = super().address_record_to_dict(record)
- if record.addressline6 == "PO1 5BZ":
+ if record.addressline6 == "PO4 099":
- return None
+ rec["postcode"] = "PO4 0PL"
- return super().address_record_to_dict(record)
+ if record.property_urn.strip().lstrip("0") in [
+ "1775122942",
+ "1775122943",
+ "1775122944",
+ ]:
+ rec["postcode"] = "PO5 2BZ"
+
+ return rec
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
- if rec["internal_council_id"] == "3270":
+ if rec["internal_council_id"] == "3596":
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec
| Add import script for Portsmouth | ## Code Before:
from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
elections = ["local.2018-05-03"]
csv_delimiter = "\t"
def address_record_to_dict(self, record):
if record.addressline6 == "PO1 5BZ":
return None
return super().address_record_to_dict(record)
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
if rec["internal_council_id"] == "3270":
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec
## Instruction:
Add import script for Portsmouth
## Code After:
from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
addresses_name = (
"local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
)
stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
elections = ["local.2019-05-02"]
csv_delimiter = "\t"
def address_record_to_dict(self, record):
rec = super().address_record_to_dict(record)
if record.addressline6 == "PO4 099":
rec["postcode"] = "PO4 0PL"
if record.property_urn.strip().lstrip("0") in [
"1775122942",
"1775122943",
"1775122944",
]:
rec["postcode"] = "PO5 2BZ"
return rec
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
if rec["internal_council_id"] == "3596":
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec
| from django.contrib.gis.geos import Point
from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter
class Command(BaseXpressDemocracyClubCsvImporter):
council_id = "E06000044"
+ addresses_name = (
- addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
? -------------- ^ ^ ^ ^ ^
+ "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
? ^^ ^ ^ ^ ^^^^^^^^^^^
+ )
- stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv"
? ^ ^ ^ ^
+ stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv"
? ^ ^ ^ ^^^^^^^^^^^
- elections = ["local.2018-05-03"]
? ^ ^
+ elections = ["local.2019-05-02"]
? ^ ^
csv_delimiter = "\t"
def address_record_to_dict(self, record):
+ rec = super().address_record_to_dict(record)
- if record.addressline6 == "PO1 5BZ":
? ^ ^^^
+ if record.addressline6 == "PO4 099":
? ^ ^^^
- return None
+ rec["postcode"] = "PO4 0PL"
- return super().address_record_to_dict(record)
+ if record.property_urn.strip().lstrip("0") in [
+ "1775122942",
+ "1775122943",
+ "1775122944",
+ ]:
+ rec["postcode"] = "PO5 2BZ"
+
+ return rec
def station_record_to_dict(self, record):
rec = super().station_record_to_dict(record)
- if rec["internal_council_id"] == "3270":
? ^^^
+ if rec["internal_council_id"] == "3596":
? ^^^
rec["location"] = Point(-1.059545, 50.7866578, srid=4326)
return rec |
0d135a746dd79ad1b703570e2bb3b27a694c67b0 | simuvex/procedures/stubs/NoReturnUnconstrained.py | simuvex/procedures/stubs/NoReturnUnconstrained.py | import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'}
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
NO_RET = True
def run(self): #pylint:disable=unused-argument
return
| import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp',
'__longjmp_chk', '__siglongjmp_chk'}
NO_RET = True
def run(self, **kwargs): #pylint:disable=unused-argument
return
| Move use_cases into the class | Move use_cases into the class
| Python | bsd-2-clause | iamahuman/angr,iamahuman/angr,f-prettyland/angr,chubbymaggie/simuvex,chubbymaggie/angr,schieb/angr,angr/angr,tyb0807/angr,f-prettyland/angr,tyb0807/angr,axt/angr,tyb0807/angr,chubbymaggie/angr,angr/angr,angr/angr,iamahuman/angr,axt/angr,f-prettyland/angr,chubbymaggie/angr,angr/simuvex,chubbymaggie/simuvex,zhuyue1314/simuvex,axt/angr,schieb/angr,chubbymaggie/simuvex,schieb/angr | import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
- use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'}
-
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
+ use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp',
+ '__longjmp_chk', '__siglongjmp_chk'}
NO_RET = True
- def run(self): #pylint:disable=unused-argument
+ def run(self, **kwargs): #pylint:disable=unused-argument
return
| Move use_cases into the class | ## Code Before:
import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'}
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
NO_RET = True
def run(self): #pylint:disable=unused-argument
return
## Instruction:
Move use_cases into the class
## Code After:
import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp',
'__longjmp_chk', '__siglongjmp_chk'}
NO_RET = True
def run(self, **kwargs): #pylint:disable=unused-argument
return
| import simuvex
######################################
# NoReturnUnconstrained
# Use in places you would put ReturnUnconstrained as a default action
# But the function shouldn't actually return
######################################
- use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'}
-
class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin
+ use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp',
+ '__longjmp_chk', '__siglongjmp_chk'}
NO_RET = True
- def run(self): #pylint:disable=unused-argument
+ def run(self, **kwargs): #pylint:disable=unused-argument
? ++++++++++
return |
27a33628310cbd68632f0e8b514de731a033f8e6 | IPython/utils/tests/test_shimmodule.py | IPython/utils/tests/test_shimmodule.py | import sys
import warnings
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("always")
import IPython.config
assert len(w) == 1
assert issubclass(w[-1].category, ShimWarning)
| import pytest
import sys
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
with pytest.warns(ShimWarning):
import IPython.config
| Make test_shim_warning not fail on unrelated warnings | Make test_shim_warning not fail on unrelated warnings
| Python | bsd-3-clause | ipython/ipython,ipython/ipython | + import pytest
import sys
- import warnings
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
+ with pytest.warns(ShimWarning):
- with warnings.catch_warnings(record=True) as w:
- warnings.simplefilter("always")
import IPython.config
- assert len(w) == 1
- assert issubclass(w[-1].category, ShimWarning)
| Make test_shim_warning not fail on unrelated warnings | ## Code Before:
import sys
import warnings
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
with warnings.catch_warnings(record=True) as w:
warnings.simplefilter("always")
import IPython.config
assert len(w) == 1
assert issubclass(w[-1].category, ShimWarning)
## Instruction:
Make test_shim_warning not fail on unrelated warnings
## Code After:
import pytest
import sys
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
with pytest.warns(ShimWarning):
import IPython.config
| + import pytest
import sys
- import warnings
from IPython.utils.shimmodule import ShimWarning
def test_shim_warning():
sys.modules.pop('IPython.config', None)
+ with pytest.warns(ShimWarning):
- with warnings.catch_warnings(record=True) as w:
- warnings.simplefilter("always")
import IPython.config
- assert len(w) == 1
- assert issubclass(w[-1].category, ShimWarning) |
74816d4af07808009b89163060f97014b1a20ceb | tests/test_arguments.py | tests/test_arguments.py | import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
'__cmp__', '__hash__']
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv'
| import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
'__cmp__', '__hash__', '__nonzero__']
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv'
| Enforce that arguments must implement non-zero methods. | Enforce that arguments must implement non-zero methods. | Python | apache-2.0 | disqus/gutter,disqus/gutter,kalail/gutter,kalail/gutter,kalail/gutter | import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
- '__cmp__', '__hash__']
+ '__cmp__', '__hash__', '__nonzero__']
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv'
| Enforce that arguments must implement non-zero methods. | ## Code Before:
import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
'__cmp__', '__hash__']
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv'
## Instruction:
Enforce that arguments must implement non-zero methods.
## Code After:
import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
'__cmp__', '__hash__', '__nonzero__']
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv'
| import unittest
from mock import MagicMock, Mock
from nose.tools import *
from gargoyle.inputs.arguments import *
class BaseArgument(object):
def setUp(self):
self.argument = self.klass(self.valid_comparison_value)
@property
def interface_functions(self):
return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__',
- '__cmp__', '__hash__']
+ '__cmp__', '__hash__', '__nonzero__']
? +++++++++++++++
@property
def interface_methods(self):
return [getattr(self.argument, f) for f in self.interface_functions]
def test_implements_comparison_methods(self):
map(ok_, self.interface_methods)
class DelegateToValue(object):
def test_delegates_all_interface_function_to_the_value_passed_in(self):
value_passed_in = MagicMock()
value_passed_in.__cmp__ = Mock()
argument = self.klass(value_passed_in)
for function in self.interface_functions:
values_function = getattr(value_passed_in, function)
arguments_function = getattr(argument, function)
arguments_function(self.valid_comparison_value)
values_function.assert_called_once_with(self.valid_comparison_value)
class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase):
klass = Value
@property
def valid_comparison_value(self):
return 'marv' |
61a4743b62914559fea18a945f7a780e1394da2f | test/test_export_flow.py | test/test_export_flow.py | import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
headers=None,
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
def test_request_simple():
flow = tutils.tflow(req=req_get)
assert flow_export.curl_command(flow)
flow = tutils.tflow(req=req_post)
assert flow_export.curl_command(flow)
| import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
req_patch = netlib.tutils.treq(
method='PATCH',
path=b"/path?query=param",
)
def test_curl_command():
flow = tutils.tflow(req=req_get)
result = """curl -H 'header:qvalue' 'http://address/path'"""
assert flow_export.curl_command(flow) == result
flow = tutils.tflow(req=req_post)
result = """curl -X POST 'http://address/path' --data-binary 'content'"""
assert flow_export.curl_command(flow) == result
flow = tutils.tflow(req=req_patch)
result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'"""
assert flow_export.curl_command(flow) == result
| Test exact return value of flow_export.curl_command | Test exact return value of flow_export.curl_command
| Python | mit | jvillacorta/mitmproxy,tdickers/mitmproxy,ddworken/mitmproxy,StevenVanAcker/mitmproxy,cortesi/mitmproxy,vhaupert/mitmproxy,tdickers/mitmproxy,mosajjal/mitmproxy,mosajjal/mitmproxy,fimad/mitmproxy,fimad/mitmproxy,ujjwal96/mitmproxy,vhaupert/mitmproxy,dwfreed/mitmproxy,ParthGanatra/mitmproxy,xaxa89/mitmproxy,mhils/mitmproxy,mhils/mitmproxy,tdickers/mitmproxy,StevenVanAcker/mitmproxy,mitmproxy/mitmproxy,ikoz/mitmproxy,dufferzafar/mitmproxy,mitmproxy/mitmproxy,tdickers/mitmproxy,vhaupert/mitmproxy,StevenVanAcker/mitmproxy,Kriechi/mitmproxy,zlorb/mitmproxy,gzzhanghao/mitmproxy,ikoz/mitmproxy,MatthewShao/mitmproxy,mhils/mitmproxy,mosajjal/mitmproxy,gzzhanghao/mitmproxy,mitmproxy/mitmproxy,xaxa89/mitmproxy,cortesi/mitmproxy,ujjwal96/mitmproxy,jvillacorta/mitmproxy,ddworken/mitmproxy,MatthewShao/mitmproxy,cortesi/mitmproxy,laurmurclar/mitmproxy,zlorb/mitmproxy,dwfreed/mitmproxy,dwfreed/mitmproxy,laurmurclar/mitmproxy,dwfreed/mitmproxy,mitmproxy/mitmproxy,mosajjal/mitmproxy,dufferzafar/mitmproxy,Kriechi/mitmproxy,dufferzafar/mitmproxy,ikoz/mitmproxy,jvillacorta/mitmproxy,ikoz/mitmproxy,mhils/mitmproxy,MatthewShao/mitmproxy,gzzhanghao/mitmproxy,ujjwal96/mitmproxy,ujjwal96/mitmproxy,gzzhanghao/mitmproxy,cortesi/mitmproxy,laurmurclar/mitmproxy,ParthGanatra/mitmproxy,mitmproxy/mitmproxy,zlorb/mitmproxy,fimad/mitmproxy,zlorb/mitmproxy,xaxa89/mitmproxy,mhils/mitmproxy,StevenVanAcker/mitmproxy,Kriechi/mitmproxy,vhaupert/mitmproxy,laurmurclar/mitmproxy,ParthGanatra/mitmproxy,fimad/mitmproxy,Kriechi/mitmproxy,dufferzafar/mitmproxy,ddworken/mitmproxy,ParthGanatra/mitmproxy,jvillacorta/mitmproxy,ddworken/mitmproxy,xaxa89/mitmproxy,MatthewShao/mitmproxy | import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
- headers=None,
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
+ req_patch = netlib.tutils.treq(
+ method='PATCH',
+ path=b"/path?query=param",
+ )
- def test_request_simple():
+
+ def test_curl_command():
flow = tutils.tflow(req=req_get)
+ result = """curl -H 'header:qvalue' 'http://address/path'"""
- assert flow_export.curl_command(flow)
+ assert flow_export.curl_command(flow) == result
flow = tutils.tflow(req=req_post)
+ result = """curl -X POST 'http://address/path' --data-binary 'content'"""
- assert flow_export.curl_command(flow)
+ assert flow_export.curl_command(flow) == result
+ flow = tutils.tflow(req=req_patch)
+ result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'"""
+ assert flow_export.curl_command(flow) == result
+
+ | Test exact return value of flow_export.curl_command | ## Code Before:
import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
headers=None,
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
def test_request_simple():
flow = tutils.tflow(req=req_get)
assert flow_export.curl_command(flow)
flow = tutils.tflow(req=req_post)
assert flow_export.curl_command(flow)
## Instruction:
Test exact return value of flow_export.curl_command
## Code After:
import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
req_patch = netlib.tutils.treq(
method='PATCH',
path=b"/path?query=param",
)
def test_curl_command():
flow = tutils.tflow(req=req_get)
result = """curl -H 'header:qvalue' 'http://address/path'"""
assert flow_export.curl_command(flow) == result
flow = tutils.tflow(req=req_post)
result = """curl -X POST 'http://address/path' --data-binary 'content'"""
assert flow_export.curl_command(flow) == result
flow = tutils.tflow(req=req_patch)
result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'"""
assert flow_export.curl_command(flow) == result
| import netlib.tutils
from libmproxy import flow_export
from . import tutils
req_get = netlib.tutils.treq(
method='GET',
- headers=None,
content=None,
)
req_post = netlib.tutils.treq(
method='POST',
headers=None,
)
+ req_patch = netlib.tutils.treq(
+ method='PATCH',
+ path=b"/path?query=param",
+ )
- def test_request_simple():
+
+ def test_curl_command():
flow = tutils.tflow(req=req_get)
+ result = """curl -H 'header:qvalue' 'http://address/path'"""
- assert flow_export.curl_command(flow)
+ assert flow_export.curl_command(flow) == result
? ++++++++++
flow = tutils.tflow(req=req_post)
+ result = """curl -X POST 'http://address/path' --data-binary 'content'"""
- assert flow_export.curl_command(flow)
+ assert flow_export.curl_command(flow) == result
? ++++++++++
+
+ flow = tutils.tflow(req=req_patch)
+ result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'"""
+ assert flow_export.curl_command(flow) == result
+ |
ee80818b8ff12cd351581b4c1652e64561d34a4c | rest_framework_simplejwt/token_blacklist/models.py | rest_framework_simplejwt/token_blacklist/models.py | from django.contrib.auth import get_user_model
from django.db import models
from django.utils.six import python_2_unicode_compatible
User = get_user_model()
@python_2_unicode_compatible
class OutstandingToken(models.Model):
user = models.ForeignKey(User, on_delete=models.CASCADE)
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user)
| from django.conf import settings
from django.db import models
from django.utils.six import python_2_unicode_compatible
@python_2_unicode_compatible
class OutstandingToken(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user)
| Fix broken tests in 1.8-1.10 | Fix broken tests in 1.8-1.10
| Python | mit | davesque/django-rest-framework-simplejwt,davesque/django-rest-framework-simplejwt | - from django.contrib.auth import get_user_model
+ from django.conf import settings
from django.db import models
from django.utils.six import python_2_unicode_compatible
- User = get_user_model()
-
-
@python_2_unicode_compatible
class OutstandingToken(models.Model):
- user = models.ForeignKey(User, on_delete=models.CASCADE)
+ user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user)
| Fix broken tests in 1.8-1.10 | ## Code Before:
from django.contrib.auth import get_user_model
from django.db import models
from django.utils.six import python_2_unicode_compatible
User = get_user_model()
@python_2_unicode_compatible
class OutstandingToken(models.Model):
user = models.ForeignKey(User, on_delete=models.CASCADE)
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user)
## Instruction:
Fix broken tests in 1.8-1.10
## Code After:
from django.conf import settings
from django.db import models
from django.utils.six import python_2_unicode_compatible
@python_2_unicode_compatible
class OutstandingToken(models.Model):
user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user)
| - from django.contrib.auth import get_user_model
+ from django.conf import settings
from django.db import models
from django.utils.six import python_2_unicode_compatible
- User = get_user_model()
-
-
@python_2_unicode_compatible
class OutstandingToken(models.Model):
- user = models.ForeignKey(User, on_delete=models.CASCADE)
? - ^
+ user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)
? ^^^^^^^^^^^^^^^^^^^^^^
jti = models.UUIDField(unique=True)
token = models.TextField()
created_at = models.DateTimeField()
expires_at = models.DateTimeField()
class Meta:
ordering = ('user',)
def __str__(self):
return 'Token for {} ({})'.format(
self.user,
self.jti,
)
@python_2_unicode_compatible
class BlacklistedToken(models.Model):
token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE)
blacklisted_at = models.DateTimeField(auto_now_add=True)
def __str__(self):
return 'Blacklisted token for {}'.format(self.token.user) |
5ac8e4619473275f2f0b26b8a9b64049d793a4ed | rmqid/__init__.py | rmqid/__init__.py | __version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
| __version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
import logging
try:
from logging import NullHandler
except ImportError:
# Python 2.6 does not have a NullHandler
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger().addHandler(NullHandler())
| Add a NullHandler so logging warnings are not emitted if no logger is setup | Add a NullHandler so logging warnings are not emitted if no logger is setup
| Python | bsd-3-clause | jonahbull/rabbitpy,gmr/rabbitpy,gmr/rabbitpy | __version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
+ import logging
+
+ try:
+ from logging import NullHandler
+ except ImportError:
+ # Python 2.6 does not have a NullHandler
+ class NullHandler(logging.Handler):
+ def emit(self, record):
+ pass
+
+ logging.getLogger().addHandler(NullHandler())
+ | Add a NullHandler so logging warnings are not emitted if no logger is setup | ## Code Before:
__version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
## Instruction:
Add a NullHandler so logging warnings are not emitted if no logger is setup
## Code After:
__version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
import logging
try:
from logging import NullHandler
except ImportError:
# Python 2.6 does not have a NullHandler
class NullHandler(logging.Handler):
def emit(self, record):
pass
logging.getLogger().addHandler(NullHandler())
| __version__ = '0.3.0'
from rmqid.connection import Connection
from rmqid.exchange import Exchange
from rmqid.message import Message
from rmqid.queue import Queue
from rmqid.tx import Tx
from rmqid.simple import consumer
from rmqid.simple import get
from rmqid.simple import publish
+
+ import logging
+
+ try:
+ from logging import NullHandler
+ except ImportError:
+ # Python 2.6 does not have a NullHandler
+ class NullHandler(logging.Handler):
+ def emit(self, record):
+ pass
+
+ logging.getLogger().addHandler(NullHandler()) |
257e8d2e6d1dc3c10eb7fc26c3deacaf4133bd9b | enactiveagents/view/agentevents.py | enactiveagents/view/agentevents.py |
import events
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
if event.agent not in self.preparation_history:
self.preparation_history[event.agent] = []
self.preparation_history[event.agent].append(event.action)
elif isinstance(event, events.AgentEnactionEvent):
if event.agent not in self.enaction_history:
self.enaction_history[event.agent] = []
self.enaction_history[event.agent].append(event.action)
elif isinstance(event, events.TickEvent):
pass |
import events
import json
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
if str(event.agent) not in self.preparation_history:
self.preparation_history[str(event.agent)] = []
self.preparation_history[str(event.agent)].append(str(event.action))
if len(self.preparation_history) > 20:
self.preparation_history.pop(0)
elif isinstance(event, events.AgentEnactionEvent):
if str(event.agent) not in self.enaction_history:
self.enaction_history[str(event.agent)] = []
self.enaction_history[str(event.agent)].append(str(event.action))
if len(self.enaction_history) > 20:
self.enaction_history.pop(0)
elif isinstance(event, events.TickEvent):
self.write_to_file()
def write_to_file(self):
"""
Write the history to the traces file.
"""
d = dict()
d["preparation_history"] = self.preparation_history
d["enaction_history"] = self.enaction_history
with open(self.file_path,'w+') as f:
json.dump(d, f) | Write agent events to a traces history file for the website. | Write agent events to a traces history file for the website.
| Python | mit | Beskhue/enactive-agents,Beskhue/enactive-agents,Beskhue/enactive-agents |
import events
+ import json
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
- if event.agent not in self.preparation_history:
+ if str(event.agent) not in self.preparation_history:
- self.preparation_history[event.agent] = []
+ self.preparation_history[str(event.agent)] = []
- self.preparation_history[event.agent].append(event.action)
+ self.preparation_history[str(event.agent)].append(str(event.action))
+
+ if len(self.preparation_history) > 20:
+ self.preparation_history.pop(0)
elif isinstance(event, events.AgentEnactionEvent):
- if event.agent not in self.enaction_history:
+ if str(event.agent) not in self.enaction_history:
- self.enaction_history[event.agent] = []
+ self.enaction_history[str(event.agent)] = []
- self.enaction_history[event.agent].append(event.action)
+ self.enaction_history[str(event.agent)].append(str(event.action))
+
+ if len(self.enaction_history) > 20:
+ self.enaction_history.pop(0)
elif isinstance(event, events.TickEvent):
- pass
+ self.write_to_file()
+
+ def write_to_file(self):
+ """
+ Write the history to the traces file.
+ """
+
+ d = dict()
+ d["preparation_history"] = self.preparation_history
+ d["enaction_history"] = self.enaction_history
+ with open(self.file_path,'w+') as f:
+ json.dump(d, f) | Write agent events to a traces history file for the website. | ## Code Before:
import events
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
if event.agent not in self.preparation_history:
self.preparation_history[event.agent] = []
self.preparation_history[event.agent].append(event.action)
elif isinstance(event, events.AgentEnactionEvent):
if event.agent not in self.enaction_history:
self.enaction_history[event.agent] = []
self.enaction_history[event.agent].append(event.action)
elif isinstance(event, events.TickEvent):
pass
## Instruction:
Write agent events to a traces history file for the website.
## Code After:
import events
import json
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
if str(event.agent) not in self.preparation_history:
self.preparation_history[str(event.agent)] = []
self.preparation_history[str(event.agent)].append(str(event.action))
if len(self.preparation_history) > 20:
self.preparation_history.pop(0)
elif isinstance(event, events.AgentEnactionEvent):
if str(event.agent) not in self.enaction_history:
self.enaction_history[str(event.agent)] = []
self.enaction_history[str(event.agent)].append(str(event.action))
if len(self.enaction_history) > 20:
self.enaction_history.pop(0)
elif isinstance(event, events.TickEvent):
self.write_to_file()
def write_to_file(self):
"""
Write the history to the traces file.
"""
d = dict()
d["preparation_history"] = self.preparation_history
d["enaction_history"] = self.enaction_history
with open(self.file_path,'w+') as f:
json.dump(d, f) |
import events
+ import json
class AgentEvents(events.EventListener):
"""
View class
"""
def __init__(self, file_path):
"""
:param file_path: The path of the file to output the history to.
"""
self.file_path = file_path
self.preparation_history = dict()
self.enaction_history = dict()
def notify(self, event):
if isinstance(event, events.AgentPreparationEvent):
- if event.agent not in self.preparation_history:
+ if str(event.agent) not in self.preparation_history:
? ++++ +
- self.preparation_history[event.agent] = []
+ self.preparation_history[str(event.agent)] = []
? ++++ +
- self.preparation_history[event.agent].append(event.action)
+ self.preparation_history[str(event.agent)].append(str(event.action))
? ++++ + ++++ +
+
+ if len(self.preparation_history) > 20:
+ self.preparation_history.pop(0)
elif isinstance(event, events.AgentEnactionEvent):
- if event.agent not in self.enaction_history:
+ if str(event.agent) not in self.enaction_history:
? ++++ +
- self.enaction_history[event.agent] = []
+ self.enaction_history[str(event.agent)] = []
? ++++ +
- self.enaction_history[event.agent].append(event.action)
+ self.enaction_history[str(event.agent)].append(str(event.action))
? ++++ + ++++ +
+
+ if len(self.enaction_history) > 20:
+ self.enaction_history.pop(0)
elif isinstance(event, events.TickEvent):
- pass
+ self.write_to_file()
+
+ def write_to_file(self):
+ """
+ Write the history to the traces file.
+ """
+
+ d = dict()
+ d["preparation_history"] = self.preparation_history
+ d["enaction_history"] = self.enaction_history
+ with open(self.file_path,'w+') as f:
+ json.dump(d, f) |
a24d6a25cb7ee5101e8131a9719744f79b23c11b | examples/quotes/quotes.py | examples/quotes/quotes.py | import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
quote = random.choice(quotes)
nw0.send_notification(address, "quote", quote)
time.sleep(0.5)
if __name__ == '__main__':
main(*sys.argv[1:])
| import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
quote = random.choice(quotes)
nw0.send_notification(my_address, "quote", quote)
time.sleep(1)
if __name__ == '__main__':
main(*sys.argv[1:])
| Send notification to the correct address | Send notification to the correct address
| Python | mit | tjguk/networkzero,tjguk/networkzero,tjguk/networkzero | import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
-
+
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
- quote = random.choice(quotes)
- nw0.send_notification(address, "quote", quote)
+ quote = random.choice(quotes)
+ nw0.send_notification(my_address, "quote", quote)
+
- time.sleep(0.5)
+ time.sleep(1)
if __name__ == '__main__':
main(*sys.argv[1:])
| Send notification to the correct address | ## Code Before:
import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
quote = random.choice(quotes)
nw0.send_notification(address, "quote", quote)
time.sleep(0.5)
if __name__ == '__main__':
main(*sys.argv[1:])
## Instruction:
Send notification to the correct address
## Code After:
import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
quote = random.choice(quotes)
nw0.send_notification(my_address, "quote", quote)
time.sleep(1)
if __name__ == '__main__':
main(*sys.argv[1:])
| import sys
print(sys.version_info)
import random
import time
import networkzero as nw0
quotes = [
"Humpty Dumpty sat on a wall",
"Hickory Dickory Dock",
"Baa Baa Black Sheep",
"Old King Cole was a merry old sould",
]
def main(address_pattern=None):
my_name = input("Name: ")
my_address = nw0.advertise(my_name, address_pattern)
print("Advertising %s on %s" % (my_name, my_address))
while True:
services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]
-
+
for name, address in services:
topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0)
if topic:
print("%s says: %s" % (name, message))
- quote = random.choice(quotes)
- nw0.send_notification(address, "quote", quote)
+ quote = random.choice(quotes)
+ nw0.send_notification(my_address, "quote", quote)
+
- time.sleep(0.5)
? ^^^
+ time.sleep(1)
? ^
if __name__ == '__main__':
main(*sys.argv[1:]) |
ffde5305a2182e566384887d51e4fde90adc9908 | runtests.py | runtests.py | import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
failures = test_runner.run_tests(["tests"])
sys.exit(bool(failures))
| import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
tests = "tests" if len(sys.argv) == 1 else sys.argv[1]
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
failures = test_runner.run_tests([tests])
sys.exit(bool(failures))
| Make it possible to run individual tests. | Tests: Make it possible to run individual tests.
| Python | agpl-3.0 | etesync/journal-manager | import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
+ tests = "tests" if len(sys.argv) == 1 else sys.argv[1]
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
- failures = test_runner.run_tests(["tests"])
+ failures = test_runner.run_tests([tests])
sys.exit(bool(failures))
| Make it possible to run individual tests. | ## Code Before:
import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
failures = test_runner.run_tests(["tests"])
sys.exit(bool(failures))
## Instruction:
Make it possible to run individual tests.
## Code After:
import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
tests = "tests" if len(sys.argv) == 1 else sys.argv[1]
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
failures = test_runner.run_tests([tests])
sys.exit(bool(failures))
| import os
import sys
import django
from django.conf import settings
from django.test.utils import get_runner
if __name__ == "__main__":
+ tests = "tests" if len(sys.argv) == 1 else sys.argv[1]
os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings'
django.setup()
TestRunner = get_runner(settings)
test_runner = TestRunner()
- failures = test_runner.run_tests(["tests"])
? - -
+ failures = test_runner.run_tests([tests])
sys.exit(bool(failures)) |
09d356f7b124368ac2ca80efa981d115ea847196 | django_ethereum_events/web3_service.py | django_ethereum_events/web3_service.py | from django.conf import settings
from web3 import Web3, RPCProvider
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
rpc_provider = RPCProvider(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
ssl=settings.ETHEREUM_NODE_SSL,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__()
| from django.conf import settings
from web3 import Web3
try:
from web3 import HTTPProvider
RPCProvider = None
except ImportError:
from web3 import RPCProvider
HTTPProvider = None
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
if HTTPProvider is not None:
uri = "{scheme}://{host}:{port}".format(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
scheme="https" if settings.ETHEREUM_NODE_SSL else "http",
)
rpc_provider = HTTPProvider(
endpoint_uri=uri,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
elif RPCProvider is not None:
rpc_provider = RPCProvider(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
ssl=settings.ETHEREUM_NODE_SSL,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
else:
raise ValueError("Cannot instantiate any RPC provider")
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__()
| Support for Web3 4.0beta: HTTPProvider | Support for Web3 4.0beta: HTTPProvider
In Web3 3.16 the class is called RPCProvider, but in the
upcoming 4.0 series it's replaced with HTTPProvider.
This commit ensures both versions are supported in this regard.
| Python | mit | artemistomaras/django-ethereum-events,artemistomaras/django-ethereum-events | from django.conf import settings
+ from web3 import Web3
+ try:
+ from web3 import HTTPProvider
+ RPCProvider = None
+ except ImportError:
- from web3 import Web3, RPCProvider
+ from web3 import RPCProvider
+ HTTPProvider = None
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
- rpc_provider = RPCProvider(
+ if HTTPProvider is not None:
+ uri = "{scheme}://{host}:{port}".format(
- host=settings.ETHEREUM_NODE_HOST,
+ host=settings.ETHEREUM_NODE_HOST,
- port=settings.ETHEREUM_NODE_PORT,
+ port=settings.ETHEREUM_NODE_PORT,
- ssl=settings.ETHEREUM_NODE_SSL,
+ scheme="https" if settings.ETHEREUM_NODE_SSL else "http",
+ )
+ rpc_provider = HTTPProvider(
+ endpoint_uri=uri,
- timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
+ timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
- )
+ )
+ elif RPCProvider is not None:
+ rpc_provider = RPCProvider(
+ host=settings.ETHEREUM_NODE_HOST,
+ port=settings.ETHEREUM_NODE_PORT,
+ ssl=settings.ETHEREUM_NODE_SSL,
+ timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
+ )
+ else:
+ raise ValueError("Cannot instantiate any RPC provider")
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__()
| Support for Web3 4.0beta: HTTPProvider | ## Code Before:
from django.conf import settings
from web3 import Web3, RPCProvider
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
rpc_provider = RPCProvider(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
ssl=settings.ETHEREUM_NODE_SSL,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__()
## Instruction:
Support for Web3 4.0beta: HTTPProvider
## Code After:
from django.conf import settings
from web3 import Web3
try:
from web3 import HTTPProvider
RPCProvider = None
except ImportError:
from web3 import RPCProvider
HTTPProvider = None
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
if HTTPProvider is not None:
uri = "{scheme}://{host}:{port}".format(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
scheme="https" if settings.ETHEREUM_NODE_SSL else "http",
)
rpc_provider = HTTPProvider(
endpoint_uri=uri,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
elif RPCProvider is not None:
rpc_provider = RPCProvider(
host=settings.ETHEREUM_NODE_HOST,
port=settings.ETHEREUM_NODE_PORT,
ssl=settings.ETHEREUM_NODE_SSL,
timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
)
else:
raise ValueError("Cannot instantiate any RPC provider")
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__()
| from django.conf import settings
+ from web3 import Web3
+ try:
+ from web3 import HTTPProvider
+ RPCProvider = None
+ except ImportError:
- from web3 import Web3, RPCProvider
? ------
+ from web3 import RPCProvider
? ++++
+ HTTPProvider = None
from .singleton import Singleton
class Web3Service(metaclass=Singleton):
"""Creates a `web3` instance based on the given `RPCProvider`."""
def __init__(self, *args, **kwargs):
"""Initializes the `web3` object.
Args:
rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance.
"""
rpc_provider = kwargs.pop('rpc_provider', None)
if not rpc_provider:
- rpc_provider = RPCProvider(
+ if HTTPProvider is not None:
+ uri = "{scheme}://{host}:{port}".format(
- host=settings.ETHEREUM_NODE_HOST,
+ host=settings.ETHEREUM_NODE_HOST,
? ++++
- port=settings.ETHEREUM_NODE_PORT,
+ port=settings.ETHEREUM_NODE_PORT,
? ++++
- ssl=settings.ETHEREUM_NODE_SSL,
+ scheme="https" if settings.ETHEREUM_NODE_SSL else "http",
+ )
+ rpc_provider = HTTPProvider(
+ endpoint_uri=uri,
- timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
+ timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
? ++++
- )
+ )
? ++++
+ elif RPCProvider is not None:
+ rpc_provider = RPCProvider(
+ host=settings.ETHEREUM_NODE_HOST,
+ port=settings.ETHEREUM_NODE_PORT,
+ ssl=settings.ETHEREUM_NODE_SSL,
+ timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10)
+ )
+ else:
+ raise ValueError("Cannot instantiate any RPC provider")
self.web3 = Web3(rpc_provider)
super(Web3Service, self).__init__() |
d3cb08d45af60aaf06757ad230a2a33bc3615543 | apps/organizations/middleware.py | apps/organizations/middleware.py | from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404
| from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
if request.subdomain is None:
return
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404
| Remove subdomain check on pages where subdomain is none | Remove subdomain check on pages where subdomain is none
| Python | mit | xobb1t/ddash2013,xobb1t/ddash2013 | from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
+ if request.subdomain is None:
+ return
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404
| Remove subdomain check on pages where subdomain is none | ## Code Before:
from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404
## Instruction:
Remove subdomain check on pages where subdomain is none
## Code After:
from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
if request.subdomain is None:
return
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404
| from django.http import Http404
from .models import Organization
class OrganizationMiddleware(object):
def process_request(self, request):
+ if request.subdomain is None:
+ return
try:
request.organization = Organization.objects.get(
slug__iexact=request.subdomain
)
except Organization.DoesNotExist:
raise Http404 |
b7a0653cdb2c20def38a687963763b75455ebbcb | conftest.py | conftest.py | from __future__ import absolute_import, division, print_function
from dials.conftest import regression_data, run_in_tmpdir
| from __future__ import absolute_import, division, print_function
from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
| Add --regression command line option | Add --regression command line option
| Python | bsd-3-clause | xia2/i19 | from __future__ import absolute_import, division, print_function
- from dials.conftest import regression_data, run_in_tmpdir
+ from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
| Add --regression command line option | ## Code Before:
from __future__ import absolute_import, division, print_function
from dials.conftest import regression_data, run_in_tmpdir
## Instruction:
Add --regression command line option
## Code After:
from __future__ import absolute_import, division, print_function
from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
| from __future__ import absolute_import, division, print_function
- from dials.conftest import regression_data, run_in_tmpdir
+ from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
? ++++++++++++++++++
|
7c3cf9e430bee4451e817ccc3d32884ed0c5f8e9 | bakeit/uploader.py | bakeit/uploader.py | try:
from urllib.request import urlopen, Request, HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
response = json.loads(e.read())
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"]
| try:
from urllib.request import urlopen, Request
from urllib.error import HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
response = json.loads(e.read().decode("utf8"))
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"]
| Fix Python3 error when decoding the response. | fix: Fix Python3 error when decoding the response.
| Python | mit | skorokithakis/bakeit | try:
- from urllib.request import urlopen, Request, HTTPError
+ from urllib.request import urlopen, Request
+ from urllib.error import HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
- response = json.loads(e.read())
+ response = json.loads(e.read().decode("utf8"))
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"]
| Fix Python3 error when decoding the response. | ## Code Before:
try:
from urllib.request import urlopen, Request, HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
response = json.loads(e.read())
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"]
## Instruction:
Fix Python3 error when decoding the response.
## Code After:
try:
from urllib.request import urlopen, Request
from urllib.error import HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
response = json.loads(e.read().decode("utf8"))
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"]
| try:
- from urllib.request import urlopen, Request, HTTPError
? -----------
+ from urllib.request import urlopen, Request
+ from urllib.error import HTTPError
except ImportError:
from urllib2 import urlopen, Request, HTTPError
import json
class PasteryUploader():
def __init__(self, api_key):
"""
Initialize an Uploader instance with the given API key.
"""
self.api_key = api_key
def upload(self, body, title="", language=None, duration=None, max_views=0):
"""
Upload the given body with the specified language type.
"""
url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key
if title:
url += "&title=%s" % title
if language:
url += "&language=%s" % language
if duration:
url += "&duration=%s" % duration
if max_views:
url += "&max_views=%s" % max_views
body = bytes(body.encode("utf8"))
req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'})
try:
response = urlopen(req)
except HTTPError as e:
- response = json.loads(e.read())
+ response = json.loads(e.read().decode("utf8"))
? ++++++++++++++ +
raise RuntimeError(response["error_msg"])
response = json.loads(response.read().decode("utf8"))
return response["url"] |
10948cd88d51383e13af0a116703984752092c6a | jenkinsapi_tests/systests/test_jenkins_matrix.py | jenkinsapi_tests/systests/test_jenkins_matrix.py | '''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
b = job.get_last_build()
while b.is_running():
time.sleep(1)
s = set()
for r in b.get_matrix_runs():
self.assertEquals(r.get_number(), b.get_number())
self.assertEquals(r.get_upstream_build(), b)
m = re.search(u'\xbb (.*) #\\d+$', r.name)
self.assertIsNotNone(m)
s.add(m.group(1))
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(b.get_master_job_name())
self.assertEqual(s, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
| '''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
build = job.get_last_build()
while build.is_running():
time.sleep(1)
set_of_groups = set()
for run in build.get_matrix_runs():
self.assertEquals(run.get_number(), build.get_number())
self.assertEquals(run.get_upstream_build(), build)
match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
self.assertIsNotNone(match_result)
set_of_groups.add(match_result.group(1))
build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(build.get_master_job_name())
self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
| Tidy up this test - still quite bad & useless. | Tidy up this test - still quite bad & useless.
| Python | mit | imsardine/jenkinsapi,salimfadhley/jenkinsapi,JohnLZeller/jenkinsapi,JohnLZeller/jenkinsapi,aerickson/jenkinsapi,domenkozar/jenkinsapi,zaro0508/jenkinsapi,imsardine/jenkinsapi,zaro0508/jenkinsapi,jduan/jenkinsapi,mistermocha/jenkinsapi,domenkozar/jenkinsapi,salimfadhley/jenkinsapi,zaro0508/jenkinsapi,mistermocha/jenkinsapi,aerickson/jenkinsapi,jduan/jenkinsapi,JohnLZeller/jenkinsapi,mistermocha/jenkinsapi,imsardine/jenkinsapi | '''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
- b = job.get_last_build()
+ build = job.get_last_build()
- while b.is_running():
+ while build.is_running():
time.sleep(1)
- s = set()
+ set_of_groups = set()
- for r in b.get_matrix_runs():
+ for run in build.get_matrix_runs():
- self.assertEquals(r.get_number(), b.get_number())
+ self.assertEquals(run.get_number(), build.get_number())
- self.assertEquals(r.get_upstream_build(), b)
+ self.assertEquals(run.get_upstream_build(), build)
- m = re.search(u'\xbb (.*) #\\d+$', r.name)
+ match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
- self.assertIsNotNone(m)
+ self.assertIsNotNone(match_result)
- s.add(m.group(1))
+ set_of_groups.add(match_result.group(1))
+ build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
- self.assertFalse(b.get_master_job_name())
+ self.assertFalse(build.get_master_job_name())
- self.assertEqual(s, set(['one', 'two', 'three']))
+ self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
| Tidy up this test - still quite bad & useless. | ## Code Before:
'''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
b = job.get_last_build()
while b.is_running():
time.sleep(1)
s = set()
for r in b.get_matrix_runs():
self.assertEquals(r.get_number(), b.get_number())
self.assertEquals(r.get_upstream_build(), b)
m = re.search(u'\xbb (.*) #\\d+$', r.name)
self.assertIsNotNone(m)
s.add(m.group(1))
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(b.get_master_job_name())
self.assertEqual(s, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
## Instruction:
Tidy up this test - still quite bad & useless.
## Code After:
'''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
build = job.get_last_build()
while build.is_running():
time.sleep(1)
set_of_groups = set()
for run in build.get_matrix_runs():
self.assertEquals(run.get_number(), build.get_number())
self.assertEquals(run.get_upstream_build(), build)
match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
self.assertIsNotNone(match_result)
set_of_groups.add(match_result.group(1))
build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(build.get_master_job_name())
self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
| '''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
- b = job.get_last_build()
+ build = job.get_last_build()
? ++++
- while b.is_running():
+ while build.is_running():
? ++++
time.sleep(1)
- s = set()
+ set_of_groups = set()
- for r in b.get_matrix_runs():
+ for run in build.get_matrix_runs():
? ++ ++++
- self.assertEquals(r.get_number(), b.get_number())
+ self.assertEquals(run.get_number(), build.get_number())
? ++ ++++
- self.assertEquals(r.get_upstream_build(), b)
+ self.assertEquals(run.get_upstream_build(), build)
? ++ ++++
- m = re.search(u'\xbb (.*) #\\d+$', r.name)
+ match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
? +++++++++++ ++
- self.assertIsNotNone(m)
+ self.assertIsNotNone(match_result)
? +++++++++++
- s.add(m.group(1))
+ set_of_groups.add(match_result.group(1))
+ build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
- self.assertFalse(b.get_master_job_name())
+ self.assertFalse(build.get_master_job_name())
? ++++
- self.assertEqual(s, set(['one', 'two', 'three']))
+ self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
? ++++++++++++
if __name__ == '__main__':
unittest.main() |
86dca4a7d3c1574af9da85e5a2f10b84d18d28c0 | blueprints/aws_backup_plans/delete.py | blueprints/aws_backup_plans/delete.py | from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
backup_plan_name=resource.name
backup_vault_name=backup_plan_name+'backup-vault'
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
try:
set_progress("Deleting the backup plan vault...")
client.delete_backup_vault(
BackupVaultName=backup_vault_name)
set_progress("Deleting the backup plan...")
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", ""
| from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
set_progress("Deleting the backup plan...")
try:
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", "" | Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working" | Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working"
| Python | apache-2.0 | CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge | from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
+
- backup_plan_name=resource.name
- backup_vault_name=backup_plan_name+'backup-vault'
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
-
+ set_progress("Deleting the backup plan...")
try:
- set_progress("Deleting the backup plan vault...")
- client.delete_backup_vault(
- BackupVaultName=backup_vault_name)
-
- set_progress("Deleting the backup plan...")
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", ""
- | Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working" | ## Code Before:
from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
backup_plan_name=resource.name
backup_vault_name=backup_plan_name+'backup-vault'
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
try:
set_progress("Deleting the backup plan vault...")
client.delete_backup_vault(
BackupVaultName=backup_vault_name)
set_progress("Deleting the backup plan...")
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", ""
## Instruction:
Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working"
## Code After:
from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
set_progress("Deleting the backup plan...")
try:
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", "" | from common.methods import set_progress
from azure.common.credentials import ServicePrincipalCredentials
from botocore.exceptions import ClientError
from resourcehandlers.aws.models import AWSHandler
import boto3
def run(job, **kwargs):
resource = kwargs.pop('resources').first()
backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value
rh_id = resource.attributes.get(field__name='aws_rh_id').value
region = resource.attributes.get(field__name='aws_region').value
rh = AWSHandler.objects.get(id=rh_id)
+
- backup_plan_name=resource.name
- backup_vault_name=backup_plan_name+'backup-vault'
set_progress("Connecting to aws backups...")
client = boto3.client('backup',
region_name=region,
aws_access_key_id=rh.serviceaccount,
aws_secret_access_key=rh.servicepasswd
)
-
+ set_progress("Deleting the backup plan...")
try:
- set_progress("Deleting the backup plan vault...")
- client.delete_backup_vault(
- BackupVaultName=backup_vault_name)
-
- set_progress("Deleting the backup plan...")
client.delete_backup_plan(BackupPlanId=backup_plan_id)
except Exception as e:
return "FAILURE", "Backup plan could not be deleted", e
return "SUCCESS", "The network security group has been succesfully deleted", "" |
c9ecacdb04f3f8df4f85057ad0d3c69df9481122 | core/utils/check_sanity.py | core/utils/check_sanity.py | import os
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
| import os
from core.utils.Executor import _convert_subprocess_cmd
import subprocess
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
def check_mirage():
try:
subprocess.check_call(
_convert_subprocess_cmd('which mirage')
)
except subprocess.CalledProcessError:
return False
else:
return True
| Check if mirage is installed | Check if mirage is installed
| Python | apache-2.0 | onyb/dune,adyasha/dune,adyasha/dune,adyasha/dune | import os
+ from core.utils.Executor import _convert_subprocess_cmd
+ import subprocess
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
+
+ def check_mirage():
+ try:
+ subprocess.check_call(
+ _convert_subprocess_cmd('which mirage')
+ )
+ except subprocess.CalledProcessError:
+ return False
+ else:
+ return True
+ | Check if mirage is installed | ## Code Before:
import os
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
## Instruction:
Check if mirage is installed
## Code After:
import os
from core.utils.Executor import _convert_subprocess_cmd
import subprocess
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
def check_mirage():
try:
subprocess.check_call(
_convert_subprocess_cmd('which mirage')
)
except subprocess.CalledProcessError:
return False
else:
return True
| import os
+ from core.utils.Executor import _convert_subprocess_cmd
+ import subprocess
from core.exceptions.Exceptions import OPAMConfigurationExeception
def check_environment() -> bool:
__opam_env__ = [
'CAML_LD_LIBRARY_PATH',
'MANPATH',
'PERL5LIB',
'OCAML_TOPLEVEL_PATH',
'PATH'
]
for var in __opam_env__:
if not os.environ.get(var, None):
raise OPAMConfigurationExeception
PATH = os.environ.get('PATH')
for path in PATH.split(':'):
if path.endswith(
os.path.join('.opam', 'system', 'bin')
):
return True
+
+
+ def check_mirage():
+ try:
+ subprocess.check_call(
+ _convert_subprocess_cmd('which mirage')
+ )
+ except subprocess.CalledProcessError:
+ return False
+ else:
+ return True |
e3b0ccb529dca19bb3882f9caad82dbd965c9ae0 | onnx/__init__.py | onnx/__init__.py | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
Takes a file-like object (has to implement fileno that returns a file descriptor)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
if isinstance(obj, str) or (sys.version_info[0] == 2 and
isinstance(obj, unicode_literals.unicode_or_str)):
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
else:
model.ParseFromString(obj.read())
return model
| from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
Takes a file-like object (has "read" function)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
if hasattr(obj, 'read') and callable(obj.read):
model.ParseFromString(obj.read())
else:
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
return model
| Fix string/file-like object detection in onnx.load | Fix string/file-like object detection in onnx.load
| Python | apache-2.0 | onnx/onnx,onnx/onnx,onnx/onnx,onnx/onnx | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
- Takes a file-like object (has to implement fileno that returns a file descriptor)
+ Takes a file-like object (has "read" function)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
- if isinstance(obj, str) or (sys.version_info[0] == 2 and
- isinstance(obj, unicode_literals.unicode_or_str)):
+ if hasattr(obj, 'read') and callable(obj.read):
+ model.ParseFromString(obj.read())
+ else:
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
- else:
- model.ParseFromString(obj.read())
return model
| Fix string/file-like object detection in onnx.load | ## Code Before:
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
Takes a file-like object (has to implement fileno that returns a file descriptor)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
if isinstance(obj, str) or (sys.version_info[0] == 2 and
isinstance(obj, unicode_literals.unicode_or_str)):
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
else:
model.ParseFromString(obj.read())
return model
## Instruction:
Fix string/file-like object detection in onnx.load
## Code After:
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
Takes a file-like object (has "read" function)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
if hasattr(obj, 'read') and callable(obj.read):
model.ParseFromString(obj.read())
else:
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
return model
| from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from .onnx_pb2 import *
import sys
def load(obj):
'''
Loads a binary protobuf that stores onnx graph
@params
- Takes a file-like object (has to implement fileno that returns a file descriptor)
+ Takes a file-like object (has "read" function)
or a string containing a file name
@return ONNX ModelProto object
'''
model = ModelProto()
- if isinstance(obj, str) or (sys.version_info[0] == 2 and
- isinstance(obj, unicode_literals.unicode_or_str)):
+ if hasattr(obj, 'read') and callable(obj.read):
+ model.ParseFromString(obj.read())
+ else:
with open(obj, 'rb') as f:
model.ParseFromString(f.read())
- else:
- model.ParseFromString(obj.read())
return model |
18818a8dfebcc44f9e8b582c15d6185f9a7a0c45 | minicms/templatetags/cms.py | minicms/templatetags/cms.py | from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
if request.path != '/':
active = None
active_len = 0
# Normalize path
path = request.path.rstrip('/') + '/'
for item in menu:
# Normalize path
url = item['url'].rstrip('/') + '/'
if path.startswith(url) and len(url) > active_len:
active = item
active_len = len(url)
if active is not None:
active['active'] = True
return {'menu': menu}
| from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
active = None
active_len = 0
# Normalize path
path = request.path.rstrip('/') + '/'
for item in menu:
# Normalize path
url = item['url'].rstrip('/') + '/'
# Root is only active if you have a "Home" link
if path != '/' and url == '/':
continue
if path.startswith(url) and len(url) > active_len:
active = item
active_len = len(url)
if active is not None:
active['active'] = True
return {'menu': menu}
| Allow "Home" to be active menu item | Allow "Home" to be active menu item
| Python | bsd-3-clause | adieu/allbuttonspressed,adieu/allbuttonspressed | from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
- if request.path != '/':
- active = None
+ active = None
- active_len = 0
+ active_len = 0
+ # Normalize path
+ path = request.path.rstrip('/') + '/'
+ for item in menu:
# Normalize path
- path = request.path.rstrip('/') + '/'
- for item in menu:
- # Normalize path
- url = item['url'].rstrip('/') + '/'
+ url = item['url'].rstrip('/') + '/'
+ # Root is only active if you have a "Home" link
+ if path != '/' and url == '/':
+ continue
- if path.startswith(url) and len(url) > active_len:
+ if path.startswith(url) and len(url) > active_len:
- active = item
+ active = item
- active_len = len(url)
+ active_len = len(url)
- if active is not None:
+ if active is not None:
- active['active'] = True
+ active['active'] = True
return {'menu': menu}
| Allow "Home" to be active menu item | ## Code Before:
from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
if request.path != '/':
active = None
active_len = 0
# Normalize path
path = request.path.rstrip('/') + '/'
for item in menu:
# Normalize path
url = item['url'].rstrip('/') + '/'
if path.startswith(url) and len(url) > active_len:
active = item
active_len = len(url)
if active is not None:
active['active'] = True
return {'menu': menu}
## Instruction:
Allow "Home" to be active menu item
## Code After:
from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
active = None
active_len = 0
# Normalize path
path = request.path.rstrip('/') + '/'
for item in menu:
# Normalize path
url = item['url'].rstrip('/') + '/'
# Root is only active if you have a "Home" link
if path != '/' and url == '/':
continue
if path.startswith(url) and len(url) > active_len:
active = item
active_len = len(url)
if active is not None:
active['active'] = True
return {'menu': menu}
| from ..models import Block
from django.template import Library
register = Library()
@register.simple_tag
def show_block(name):
try:
return Block.objects.get(name=name).content
except Block.DoesNotExist:
return ''
except Block.MultipleObjectsReturned:
return 'Error: Multiple blocks for "%s"' % name
@register.inclusion_tag('minicms/menu.html', takes_context=True)
def show_menu(context, name='menu'):
request = context['request']
menu = []
try:
for line in Block.objects.get(name=name).content.splitlines():
line = line.rstrip()
try:
title, url = line.rsplit(' ', 1)
except:
continue
menu.append({'title': title.strip(), 'url': url})
except Block.DoesNotExist:
pass
# Mark the best-matching URL as active
- if request.path != '/':
- active = None
? ----
+ active = None
- active_len = 0
? ----
+ active_len = 0
+ # Normalize path
+ path = request.path.rstrip('/') + '/'
+ for item in menu:
# Normalize path
- path = request.path.rstrip('/') + '/'
- for item in menu:
- # Normalize path
- url = item['url'].rstrip('/') + '/'
? ----
+ url = item['url'].rstrip('/') + '/'
+ # Root is only active if you have a "Home" link
+ if path != '/' and url == '/':
+ continue
- if path.startswith(url) and len(url) > active_len:
? ----
+ if path.startswith(url) and len(url) > active_len:
- active = item
? ----
+ active = item
- active_len = len(url)
? ----
+ active_len = len(url)
- if active is not None:
? ----
+ if active is not None:
- active['active'] = True
? ----
+ active['active'] = True
return {'menu': menu} |
eb33d70bfda4857fbd76616cf3bf7fb7d7feec71 | spoj/00005/palin.py | spoj/00005/palin.py | def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
for j in range(i + 1, mid):
palin[j] = '0'
for j in range(mid, n):
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k))
| def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
for j in range(i + 1, (n + 1) // 2):
palin[j] = '0'
for j in range((n + 1) // 2, n):
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k))
| Fix bug in ranges (to middle) | Fix bug in ranges (to middle)
- in SPOJ palin
Signed-off-by: Karel Ha <70f8965fdfb04f1fc0e708a55d9e822c449f57d3@gmail.com>
| Python | mit | mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming | def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
- for j in range(i + 1, mid):
+ for j in range(i + 1, (n + 1) // 2):
palin[j] = '0'
- for j in range(mid, n):
+ for j in range((n + 1) // 2, n):
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k))
| Fix bug in ranges (to middle) | ## Code Before:
def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
for j in range(i + 1, mid):
palin[j] = '0'
for j in range(mid, n):
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k))
## Instruction:
Fix bug in ranges (to middle)
## Code After:
def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
for j in range(i + 1, (n + 1) // 2):
palin[j] = '0'
for j in range((n + 1) // 2, n):
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k))
| def next_palindrome(k):
palin = list(k)
n = len(k)
mid = n // 2
# case 1: forward right
just_copy = False
for i in range(mid, n):
mirrored = n - 1 - i
if k[i] < k[mirrored]:
just_copy = True
if just_copy:
palin[i] = palin[mirrored]
# case 2: backward left
if not just_copy:
i = (n - 1) // 2
while i >= 0 and k[i] == '9':
i -= 1
if i >= 0:
palin[i] = str(int(k[i]) + 1)
- for j in range(i + 1, mid):
? ^^^
+ for j in range(i + 1, (n + 1) // 2):
? ^^^^^^^^^^^^
palin[j] = '0'
- for j in range(mid, n):
? ^^^
+ for j in range((n + 1) // 2, n):
? ^^^^^^^^^^^^
mirrored = n - 1 - j
palin[j] = palin[mirrored]
else:
# case 3: "99...9" -> "100..01"
palin = ['0'] * (n + 1)
palin[0] = palin[-1] = '1'
return ''.join(palin)
if __name__ == '__main__':
t = int(input())
for _ in range(t):
k = input()
print(next_palindrome(k)) |
aa014a472a39c12cf3141dd337ecc2ed1ea2cd55 | django_summernote/test_settings.py | django_summernote/test_settings.py | import django
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
__MIDDLEWARE__ = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
if django.VERSION < (1, 10):
MIDDLEWARE_CLASSES = __MIDDLEWARE__
else:
MIDDLEWARE = __MIDDLEWARE__
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN'
| DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
MIDDLEWARE = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN'
| Drop old django version support (1.10) | Drop old django version support (1.10)
| Python | mit | summernote/django-summernote,summernote/django-summernote,summernote/django-summernote | - import django
-
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
- __MIDDLEWARE__ = (
+ MIDDLEWARE = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
-
- if django.VERSION < (1, 10):
- MIDDLEWARE_CLASSES = __MIDDLEWARE__
- else:
- MIDDLEWARE = __MIDDLEWARE__
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN'
| Drop old django version support (1.10) | ## Code Before:
import django
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
__MIDDLEWARE__ = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
if django.VERSION < (1, 10):
MIDDLEWARE_CLASSES = __MIDDLEWARE__
else:
MIDDLEWARE = __MIDDLEWARE__
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN'
## Instruction:
Drop old django version support (1.10)
## Code After:
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
MIDDLEWARE = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN'
| - import django
-
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': 'django_summernote.db',
}
}
- __MIDDLEWARE__ = (
? -- --
+ MIDDLEWARE = (
'django.middleware.common.CommonMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.locale.LocaleMiddleware',
)
-
- if django.VERSION < (1, 10):
- MIDDLEWARE_CLASSES = __MIDDLEWARE__
- else:
- MIDDLEWARE = __MIDDLEWARE__
STATIC_URL = '/'
MEDIA_URL = '/media/'
MEDIA_ROOT = 'test_media'
SECRET_KEY = 'django_summernote'
ROOT_URLCONF = 'django_summernote.urls'
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.messages',
'django.contrib.sessions',
'django.contrib.staticfiles',
'django_summernote',
]
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
]
},
},
]
if django.VERSION >= (3, 0):
X_FRAME_OPTIONS = 'SAMEORIGIN' |
8237291e194aa900857fe382d0b8cefb7806c331 | ocradmin/ocrmodels/models.py | ocradmin/ocrmodels/models.py | from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
def __unicode__(self):
"""
String representation.
"""
return self.name
| from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| Improve unicode method. Whitespace cleanup | Improve unicode method. Whitespace cleanup
| Python | apache-2.0 | vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium | from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
-
def __unicode__(self):
"""
String representation.
"""
- return self.name
+ return "<%s: %s>" % (self.__class__.__name__, self.name)
-
+ | Improve unicode method. Whitespace cleanup | ## Code Before:
from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
def __unicode__(self):
"""
String representation.
"""
return self.name
## Instruction:
Improve unicode method. Whitespace cleanup
## Code After:
from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
def __unicode__(self):
"""
String representation.
"""
return "<%s: %s>" % (self.__class__.__name__, self.name)
| from django.db import models
from django.contrib.auth.models import User
from tagging.fields import TagField
import tagging
# OCR model, erm, model
class OcrModel(models.Model):
"""
OCR model objects.
"""
user = models.ForeignKey(User)
derived_from = models.ForeignKey("self", null=True, blank=True)
tags = TagField()
name = models.CharField(max_length=100, unique=True)
description = models.TextField(null=True, blank=True)
created_on = models.DateField(auto_now_add=True)
updated_on = models.DateField(null=True, blank=True)
public = models.BooleanField(default=True)
file = models.FileField(upload_to="models")
type = models.CharField(max_length=20,
choices=[("char", "Character"), ("lang", "Language")])
app = models.CharField(max_length=20,
choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")])
-
def __unicode__(self):
"""
String representation.
"""
- return self.name
+ return "<%s: %s>" % (self.__class__.__name__, self.name)
-
+ |
68a7fd8a444a8c568d716db11849f58ad7a9dee5 | django_pesapal/views.py | django_pesapal/views.py |
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
|
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
| Add support for further processing of the payment while maintaining compatibility | Add support for further processing of the payment while maintaining compatibility
| Python | bsd-3-clause | odero/django-pesapal,odero/django-pesapal |
- from django.core.urlresolvers import reverse_lazy
+ from django.core.urlresolvers import reverse_lazy, reverse
- from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
- from django.db.models.loading import get_model
+ from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
- url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
+ url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
- transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
+ self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
- merchant_reference = request.GET.get('pesapal_merchant_reference', '')
+ self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
- if transaction_id and merchant_reference:
+ if self.transaction_id and self.merchant_reference:
- transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
+ transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
+ pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
+ def get_redirect_url(self, *args, **kwargs):
+
+ try:
+ url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
+ except NoReverseMatch:
+ url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
+ kwargs={'merchant_reference': self.merchant_reference})
+ return url
+ | Add support for further processing of the payment while maintaining compatibility | ## Code Before:
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
## Instruction:
Add support for further processing of the payment while maintaining compatibility
## Code After:
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
- from django.core.urlresolvers import reverse_lazy
+ from django.core.urlresolvers import reverse_lazy, reverse
? +++++++++
- from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
- from django.db.models.loading import get_model
+ from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
- url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
+ url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
- transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
+ self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
? +++++
- merchant_reference = request.GET.get('pesapal_merchant_reference', '')
+ self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
? +++++
- if transaction_id and merchant_reference:
+ if self.transaction_id and self.merchant_reference:
? +++++ +++++
- transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
? ------------------------------------
+ transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
? +++++
+ pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
+
+ def get_redirect_url(self, *args, **kwargs):
+
+ try:
+ url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
+ except NoReverseMatch:
+ url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
+ kwargs={'merchant_reference': self.merchant_reference})
+ return url |
0a69133e44810dd0469555f62ec49eba120e6ecc | apps/storybase/utils.py | apps/storybase/utils.py | """Shared utility functions"""
from django.template.defaultfilters import slugify as django_slugify
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-')
| """Shared utility functions"""
from django.conf import settings
from django.template.defaultfilters import slugify as django_slugify
from django.utils.translation import ugettext as _
def get_language_name(language_code):
"""Convert a language code into its full (localized) name"""
languages = dict(settings.LANGUAGES)
return _(languages[language_code])
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-')
| Add utility function to convert a language code to a its full name | Add utility function to convert a language code to a its full name
| Python | mit | denverfoundation/storybase,denverfoundation/storybase,denverfoundation/storybase,denverfoundation/storybase | """Shared utility functions"""
+ from django.conf import settings
from django.template.defaultfilters import slugify as django_slugify
+ from django.utils.translation import ugettext as _
+
+ def get_language_name(language_code):
+ """Convert a language code into its full (localized) name"""
+ languages = dict(settings.LANGUAGES)
+ return _(languages[language_code])
+
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-')
| Add utility function to convert a language code to a its full name | ## Code Before:
"""Shared utility functions"""
from django.template.defaultfilters import slugify as django_slugify
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-')
## Instruction:
Add utility function to convert a language code to a its full name
## Code After:
"""Shared utility functions"""
from django.conf import settings
from django.template.defaultfilters import slugify as django_slugify
from django.utils.translation import ugettext as _
def get_language_name(language_code):
"""Convert a language code into its full (localized) name"""
languages = dict(settings.LANGUAGES)
return _(languages[language_code])
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-')
| """Shared utility functions"""
+ from django.conf import settings
from django.template.defaultfilters import slugify as django_slugify
+ from django.utils.translation import ugettext as _
+
+ def get_language_name(language_code):
+ """Convert a language code into its full (localized) name"""
+ languages = dict(settings.LANGUAGES)
+ return _(languages[language_code])
+
def slugify(value):
"""
Normalizes string, converts to lowercase, removes non-alpha characters,
converts spaces to hyphens, and truncates to 50 characters.
"""
slug = django_slugify(value)
slug = slug[:50]
return slug.rstrip('-') |
a58c3cbfa2c0147525e1afb355e355a9edeb22f8 | discussion/admin.py | discussion/admin.py | from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
exclude = ('user',)
extra = 1
model = Comment
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment)
| from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
extra = 1
model = Comment
raw_id_fields = ('user',)
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment)
| Add user back onto the comment inline for posts | Add user back onto the comment inline for posts
| Python | bsd-2-clause | lehins/lehins-discussion,lehins/lehins-discussion,incuna/django-discussion,incuna/django-discussion,lehins/lehins-discussion | from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
- exclude = ('user',)
extra = 1
model = Comment
+ raw_id_fields = ('user',)
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment)
| Add user back onto the comment inline for posts | ## Code Before:
from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
exclude = ('user',)
extra = 1
model = Comment
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment)
## Instruction:
Add user back onto the comment inline for posts
## Code After:
from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
extra = 1
model = Comment
raw_id_fields = ('user',)
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment)
| from django.contrib import admin
from discussion.models import Comment, Discussion, Post
class CommentInline(admin.TabularInline):
- exclude = ('user',)
extra = 1
model = Comment
+ raw_id_fields = ('user',)
class PostAdmin(admin.ModelAdmin):
inlines = (CommentInline,)
list_filter = ('discussion',)
class DiscussionAdmin(admin.ModelAdmin):
prepopulated_fields = {
'slug': ('name',)
}
admin.site.register(Discussion, DiscussionAdmin)
admin.site.register(Post, PostAdmin)
admin.site.register(Comment) |
1e5e2a236277dc9ba11f9fe4aff3279f692da3f7 | ploy/tests/conftest.py | ploy/tests/conftest.py | from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock
| from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
def content(self):
with open(self.path) as f:
return f.read()
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock
| Add convenience function to read tempdir files. | Add convenience function to read tempdir files.
| Python | bsd-3-clause | fschulze/ploy,ployground/ploy | from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
+ def content(self):
+ with open(self.path) as f:
+ return f.read()
+
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock
| Add convenience function to read tempdir files. | ## Code Before:
from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock
## Instruction:
Add convenience function to read tempdir files.
## Code After:
from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
def content(self):
with open(self.path) as f:
return f.read()
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock
| from mock import patch
import pytest
import os
import shutil
import tempfile
class Directory:
def __init__(self, directory):
self.directory = directory
def __getitem__(self, name):
path = os.path.join(self.directory, name)
assert not os.path.relpath(path, self.directory).startswith('..')
return File(path)
class File:
def __init__(self, path):
self.directory = os.path.dirname(path)
self.path = path
def fill(self, content):
if not os.path.exists(self.directory):
os.makedirs(self.directory)
with open(self.path, 'w') as f:
if isinstance(content, (list, tuple)):
content = '\n'.join(content)
f.write(content)
+ def content(self):
+ with open(self.path) as f:
+ return f.read()
+
@pytest.yield_fixture
def tempdir():
""" Returns an object for easy use of a temporary directory which is
cleaned up afterwards.
Use tempdir[filepath] to access files.
Use .fill(lines) on the returned object to write content to the file.
"""
directory = tempfile.mkdtemp()
yield Directory(directory)
shutil.rmtree(directory)
@pytest.yield_fixture
def ployconf(tempdir):
""" Returns a Configfile object which manages ploy.conf.
"""
yield tempdir['etc/ploy.conf']
@pytest.yield_fixture
def os_execvp_mock():
with patch("os.execvp") as os_execvp_mock:
yield os_execvp_mock |
7a374b19cf89421a73ea55fdbcd1b16b52327568 | dm_control/composer/initializer.py | dm_control/composer/initializer.py |
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
def initialize_episode(self, physics, random_state):
raise NotImplementedError
|
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
def __call__(self, physics, random_state):
raise NotImplementedError
| Rename `initialize_episode` --> `__call__` in `composer.Initializer` | Rename `initialize_episode` --> `__call__` in `composer.Initializer`
PiperOrigin-RevId: 234775654
| Python | apache-2.0 | deepmind/dm_control |
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
- def initialize_episode(self, physics, random_state):
+ def __call__(self, physics, random_state):
raise NotImplementedError
| Rename `initialize_episode` --> `__call__` in `composer.Initializer` | ## Code Before:
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
def initialize_episode(self, physics, random_state):
raise NotImplementedError
## Instruction:
Rename `initialize_episode` --> `__call__` in `composer.Initializer`
## Code After:
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
def __call__(self, physics, random_state):
raise NotImplementedError
|
"""Module defining the abstract initializer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import abc
import six
@six.add_metaclass(abc.ABCMeta)
class Initializer(object):
"""The abstract base class for an initializer."""
@abc.abstractmethod
- def initialize_episode(self, physics, random_state):
? ^^^^^ ^^^ ^^^^^^^
+ def __call__(self, physics, random_state):
? ^^^ ^ ^
raise NotImplementedError |
2c357a54e30eecb1d7b717be3ed774dcfecc2814 | src/stratis_cli/_actions/_stratis.py | src/stratis_cli/_actions/_stratis.py |
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
def list_stratisd_redundancy(namespace):
"""
List the stratisd redundancy designations.
"""
# pylint: disable=unused-argument
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
def list_stratisd_version(namespace):
"""
List the stratisd version.
"""
# pylint: disable=unused-argument
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
|
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
def list_stratisd_redundancy(_namespace):
"""
List the stratisd redundancy designations.
"""
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
def list_stratisd_version(_namespace):
"""
List the stratisd version.
"""
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
| Use '_' prefix instead of disabling pylint unused-argument lint | Use '_' prefix instead of disabling pylint unused-argument lint
It is more precise to mark the unused parameters this way.
Signed-off-by: mulhern <7b51bcf507bcd7afb72bf8663752c0ddbeb517f6@redhat.com>
| Python | apache-2.0 | stratis-storage/stratis-cli,stratis-storage/stratis-cli |
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
- def list_stratisd_redundancy(namespace):
+ def list_stratisd_redundancy(_namespace):
"""
List the stratisd redundancy designations.
"""
- # pylint: disable=unused-argument
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
- def list_stratisd_version(namespace):
+ def list_stratisd_version(_namespace):
"""
List the stratisd version.
"""
- # pylint: disable=unused-argument
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
| Use '_' prefix instead of disabling pylint unused-argument lint | ## Code Before:
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
def list_stratisd_redundancy(namespace):
"""
List the stratisd redundancy designations.
"""
# pylint: disable=unused-argument
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
def list_stratisd_version(namespace):
"""
List the stratisd version.
"""
# pylint: disable=unused-argument
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
## Instruction:
Use '_' prefix instead of disabling pylint unused-argument lint
## Code After:
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
def list_stratisd_redundancy(_namespace):
"""
List the stratisd redundancy designations.
"""
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
def list_stratisd_version(_namespace):
"""
List the stratisd version.
"""
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
|
from .._stratisd_constants import RedundancyCodes
from ._connection import get_object
from ._constants import TOP_OBJECT
from ._data import Manager
class StratisActions():
"""
Stratis actions.
"""
@staticmethod
- def list_stratisd_redundancy(namespace):
+ def list_stratisd_redundancy(_namespace):
? +
"""
List the stratisd redundancy designations.
"""
- # pylint: disable=unused-argument
for code in RedundancyCodes:
print("%s: %d" % (code.name, code.value))
@staticmethod
- def list_stratisd_version(namespace):
+ def list_stratisd_version(_namespace):
? +
"""
List the stratisd version.
"""
- # pylint: disable=unused-argument
print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT))) |
d4d73fe7d5e83c65d9abbf59ea14ed60eb23a83f | poem_reader.py | poem_reader.py | import argparse
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
args = argparser.parse_args()
data_root = args.dataroot
with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f:
tree = etree.parse(f)
root = tree.getroot()
print(root.tag)
| import argparse
import glob
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
args = argparser.parse_args()
data_root = args.dataroot
def read_xml_directory(path):
"""
Read XML files from path, parse them, and return them as list
"""
files = glob.glob(path + "*.xml")
xmls = []
for xmlfile in files:
with open(xmlfile, 'r') as f:
xmls.append(etree.parse(f))
return xmls
def find_by_block_id(xmls, block_id):
"""
Find an element by block_id from a list of lxml trees
"""
block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id))
for xml in xmls:
elements = block_xpath(xml)
if elements:
return elements[0]
some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/'
xmls = read_xml_directory(some_dir)
print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
| Read XML files from a directory and find textblock by id | Read XML files from a directory and find textblock by id
| Python | mit | dhh17/categories_norms_genres,dhh17/categories_norms_genres,dhh17/categories_norms_genres | import argparse
+ import glob
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
+ argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
+ args = argparser.parse_args()
- argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
-
- args = argparser.parse_args()
data_root = args.dataroot
- with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f:
- tree = etree.parse(f)
- root = tree.getroot()
- print(root.tag)
+ def read_xml_directory(path):
+ """
+ Read XML files from path, parse them, and return them as list
+ """
+ files = glob.glob(path + "*.xml")
+ xmls = []
+ for xmlfile in files:
+ with open(xmlfile, 'r') as f:
+ xmls.append(etree.parse(f))
+
+ return xmls
+
+
+ def find_by_block_id(xmls, block_id):
+ """
+ Find an element by block_id from a list of lxml trees
+ """
+ block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id))
+ for xml in xmls:
+ elements = block_xpath(xml)
+ if elements:
+ return elements[0]
+
+
+ some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/'
+ xmls = read_xml_directory(some_dir)
+
+ print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
+
+
+ | Read XML files from a directory and find textblock by id | ## Code Before:
import argparse
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
args = argparser.parse_args()
data_root = args.dataroot
with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f:
tree = etree.parse(f)
root = tree.getroot()
print(root.tag)
## Instruction:
Read XML files from a directory and find textblock by id
## Code After:
import argparse
import glob
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
args = argparser.parse_args()
data_root = args.dataroot
def read_xml_directory(path):
"""
Read XML files from path, parse them, and return them as list
"""
files = glob.glob(path + "*.xml")
xmls = []
for xmlfile in files:
with open(xmlfile, 'r') as f:
xmls.append(etree.parse(f))
return xmls
def find_by_block_id(xmls, block_id):
"""
Find an element by block_id from a list of lxml trees
"""
block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id))
for xml in xmls:
elements = block_xpath(xml)
if elements:
return elements[0]
some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/'
xmls = read_xml_directory(some_dir)
print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
| import argparse
+ import glob
from lxml import etree
argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@')
+ argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
+ args = argparser.parse_args()
- argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory")
-
- args = argparser.parse_args()
data_root = args.dataroot
- with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f:
- tree = etree.parse(f)
- root = tree.getroot()
- print(root.tag)
+ def read_xml_directory(path):
+ """
+ Read XML files from path, parse them, and return them as list
+ """
+ files = glob.glob(path + "*.xml")
+
+ xmls = []
+ for xmlfile in files:
+ with open(xmlfile, 'r') as f:
+ xmls.append(etree.parse(f))
+
+ return xmls
+
+
+ def find_by_block_id(xmls, block_id):
+ """
+ Find an element by block_id from a list of lxml trees
+ """
+ block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id))
+ for xml in xmls:
+ elements = block_xpath(xml)
+ if elements:
+ return elements[0]
+
+
+ some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/'
+ xmls = read_xml_directory(some_dir)
+
+ print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
+
+ |
b583c5fb00d1ebfa0458a6233be85d8b56173abf | python/printbag.py | python/printbag.py |
import sys
import logging
import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
if topics is None:
#topics = ['/tf', '/scan']
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> \n\n'
'Print contents of rosbag file.'
).format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
| import sys
import logging
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
def print_bag(bag, topics=None):
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} [topics] <rosbag> \n\n'
'topics:\tcomma-separated list of topics\n\n'
'Print contents of rosbag file. If topics is not provided, \n'
'all topics are printed\n'
).format(__file__))
sys.exit(1)
topics = None
if len(sys.argv) == 3:
topics = [t.strip() for t in sys.argv[1].split(',')]
filename = sys.argv[2]
else:
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag, topics)
sys.exit()
| Add argument to specify bag topics | Add argument to specify bag topics
| Python | bsd-2-clause | oliverlee/antlia | -
import sys
import logging
-
- import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
- from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
- if topics is None:
- #topics = ['/tf', '/scan']
- topics = ['/scan', '/flagbutton_pressed']
-
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
- print(('Usage: {} <rosbag> \n\n'
+ print(('Usage: {} [topics] <rosbag> \n\n'
+ 'topics:\tcomma-separated list of topics\n\n'
- 'Print contents of rosbag file.'
+ 'Print contents of rosbag file. If topics is not provided, \n'
+ 'all topics are printed\n'
).format(__file__))
sys.exit(1)
- outfile = None
+ topics = None
+ if len(sys.argv) == 3:
+ topics = [t.strip() for t in sys.argv[1].split(',')]
+ filename = sys.argv[2]
+ else:
- filename = sys.argv[1]
+ filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
- print_bag(bag)
+ print_bag(bag, topics)
sys.exit()
| Add argument to specify bag topics | ## Code Before:
import sys
import logging
import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
if topics is None:
#topics = ['/tf', '/scan']
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> \n\n'
'Print contents of rosbag file.'
).format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
## Instruction:
Add argument to specify bag topics
## Code After:
import sys
import logging
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
def print_bag(bag, topics=None):
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} [topics] <rosbag> \n\n'
'topics:\tcomma-separated list of topics\n\n'
'Print contents of rosbag file. If topics is not provided, \n'
'all topics are printed\n'
).format(__file__))
sys.exit(1)
topics = None
if len(sys.argv) == 3:
topics = [t.strip() for t in sys.argv[1].split(',')]
filename = sys.argv[2]
else:
filename = sys.argv[1]
with rosbag.Bag(filename) as bag:
print_bag(bag, topics)
sys.exit()
| -
import sys
import logging
-
- import numpy as np
# suppress logging warnings due to rospy
logging.basicConfig(filename='/dev/null')
import rosbag
- from antlia.dtype import LIDAR_CONVERTED_DTYPE
def print_bag(bag, topics=None):
- if topics is None:
- #topics = ['/tf', '/scan']
- topics = ['/scan', '/flagbutton_pressed']
-
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
- print(('Usage: {} <rosbag> \n\n'
+ print(('Usage: {} [topics] <rosbag> \n\n'
? +++++++++
+ 'topics:\tcomma-separated list of topics\n\n'
- 'Print contents of rosbag file.'
+ 'Print contents of rosbag file. If topics is not provided, \n'
? ++++++++++++++++++++++++++++++
+ 'all topics are printed\n'
).format(__file__))
sys.exit(1)
- outfile = None
+ topics = None
+ if len(sys.argv) == 3:
+ topics = [t.strip() for t in sys.argv[1].split(',')]
+ filename = sys.argv[2]
+ else:
- filename = sys.argv[1]
+ filename = sys.argv[1]
? ++++
with rosbag.Bag(filename) as bag:
- print_bag(bag)
+ print_bag(bag, topics)
? ++++++++
sys.exit() |
edf38ad11631ad5e793eb9ac95dbc865595d517b | glue_vispy_viewers/common/layer_state.py | glue_vispy_viewers/common/layer_state.py | from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
| from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
| Make sure layer artist icon updates when changing the color mode or colormaps | Make sure layer artist icon updates when changing the color mode or colormaps | Python | bsd-2-clause | glue-viz/glue-vispy-viewers,PennyQ/astro-vispy,astrofrog/glue-3d-viewer,glue-viz/glue-3d-viewer,astrofrog/glue-vispy-viewers | from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
+ from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
+ self.add_global_callback(self._notify_layer_update)
+
+ def _notify_layer_update(self, **kwargs):
+ message = LayerArtistUpdatedMessage(self)
+ if self.layer is not None and self.layer.hub is not None:
+ self.layer.hub.broadcast(message)
+
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
| Make sure layer artist icon updates when changing the color mode or colormaps | ## Code Before:
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
## Instruction:
Make sure layer artist icon updates when changing the color mode or colormaps
## Code After:
from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
self.add_global_callback(self._notify_layer_update)
def _notify_layer_update(self, **kwargs):
message = LayerArtistUpdatedMessage(self)
if self.layer is not None and self.layer.hub is not None:
self.layer.hub.broadcast(message)
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
| from __future__ import absolute_import, division, print_function
from glue.external.echo import CallbackProperty, keep_in_sync
from glue.core.state_objects import State
+ from glue.core.message import LayerArtistUpdatedMessage
__all__ = ['VispyLayerState']
class VispyLayerState(State):
"""
A base state object for all Vispy layers
"""
layer = CallbackProperty()
visible = CallbackProperty(True)
zorder = CallbackProperty(0)
color = CallbackProperty()
alpha = CallbackProperty()
def __init__(self, **kwargs):
super(VispyLayerState, self).__init__(**kwargs)
self._sync_color = None
self._sync_alpha = None
self.add_callback('layer', self._layer_changed)
self._layer_changed()
+ self.add_global_callback(self._notify_layer_update)
+
+ def _notify_layer_update(self, **kwargs):
+ message = LayerArtistUpdatedMessage(self)
+ if self.layer is not None and self.layer.hub is not None:
+ self.layer.hub.broadcast(message)
+
def _layer_changed(self):
if self._sync_color is not None:
self._sync_color.stop_syncing()
if self._sync_alpha is not None:
self._sync_alpha.stop_syncing()
if self.layer is not None:
self.color = self.layer.style.color
self.alpha = self.layer.style.alpha
self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color')
self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha') |
e836f3c558085aa0a1275546ac45b8146254ee6b | test/default.py | test/default.py | from mock import MagicMock
import pbclient
class TestDefault(object):
"""Test class for pbs.helpers."""
error = {"action": "GET",
"exception_cls": "NotFound",
"exception_msg": "(NotFound)",
"status": "failed",
"status_code": 404,
"target": "/api/app"}
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
'description': 'description',
'short_name': 'short_name'}
def tearDown(self):
self.error['status'] = 'failed'
| """Test module for pbs client."""
from mock import MagicMock
import pbclient
class TestDefault(object):
"""Test class for pbs.helpers."""
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
'description': 'description',
'short_name': 'short_name'}
def tearDown(self):
"""Tear down method."""
self.error['status'] = 'failed'
@property
def error(self, action='GET',
exception_cls='NotFound',
exception_msg='(NotFound)',
status='failed',
status_code=404,
target='/api/app'):
"""Error property."""
return {'action': action,
'exception_cls': exception_cls,
'exception_msg': exception_msg,
'status': status,
'status_code': status_code,
'target': target}
| Refactor error as a property. | Refactor error as a property.
| Python | agpl-3.0 | PyBossa/pbs,PyBossa/pbs,PyBossa/pbs | + """Test module for pbs client."""
from mock import MagicMock
import pbclient
+
+
class TestDefault(object):
"""Test class for pbs.helpers."""
-
- error = {"action": "GET",
- "exception_cls": "NotFound",
- "exception_msg": "(NotFound)",
- "status": "failed",
- "status_code": 404,
- "target": "/api/app"}
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
- 'description': 'description',
+ 'description': 'description',
- 'short_name': 'short_name'}
+ 'short_name': 'short_name'}
def tearDown(self):
+ """Tear down method."""
self.error['status'] = 'failed'
+ @property
+ def error(self, action='GET',
+ exception_cls='NotFound',
+ exception_msg='(NotFound)',
+ status='failed',
+ status_code=404,
+ target='/api/app'):
+ """Error property."""
+ return {'action': action,
+ 'exception_cls': exception_cls,
+ 'exception_msg': exception_msg,
+ 'status': status,
+ 'status_code': status_code,
+ 'target': target}
+ | Refactor error as a property. | ## Code Before:
from mock import MagicMock
import pbclient
class TestDefault(object):
"""Test class for pbs.helpers."""
error = {"action": "GET",
"exception_cls": "NotFound",
"exception_msg": "(NotFound)",
"status": "failed",
"status_code": 404,
"target": "/api/app"}
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
'description': 'description',
'short_name': 'short_name'}
def tearDown(self):
self.error['status'] = 'failed'
## Instruction:
Refactor error as a property.
## Code After:
"""Test module for pbs client."""
from mock import MagicMock
import pbclient
class TestDefault(object):
"""Test class for pbs.helpers."""
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
'description': 'description',
'short_name': 'short_name'}
def tearDown(self):
"""Tear down method."""
self.error['status'] = 'failed'
@property
def error(self, action='GET',
exception_cls='NotFound',
exception_msg='(NotFound)',
status='failed',
status_code=404,
target='/api/app'):
"""Error property."""
return {'action': action,
'exception_cls': exception_cls,
'exception_msg': exception_msg,
'status': status,
'status_code': status_code,
'target': target}
| + """Test module for pbs client."""
from mock import MagicMock
import pbclient
+
+
class TestDefault(object):
"""Test class for pbs.helpers."""
-
- error = {"action": "GET",
- "exception_cls": "NotFound",
- "exception_msg": "(NotFound)",
- "status": "failed",
- "status_code": 404,
- "target": "/api/app"}
config = MagicMock()
config.server = 'http://server'
config.api_key = 'apikey'
config.pbclient = pbclient
config.project = {'name': 'name',
- 'description': 'description',
+ 'description': 'description',
? +++++
- 'short_name': 'short_name'}
+ 'short_name': 'short_name'}
? +++++
def tearDown(self):
+ """Tear down method."""
self.error['status'] = 'failed'
+
+ @property
+ def error(self, action='GET',
+ exception_cls='NotFound',
+ exception_msg='(NotFound)',
+ status='failed',
+ status_code=404,
+ target='/api/app'):
+ """Error property."""
+ return {'action': action,
+ 'exception_cls': exception_cls,
+ 'exception_msg': exception_msg,
+ 'status': status,
+ 'status_code': status_code,
+ 'target': target} |
fbc5e2d52549452c2adbe58644358cf3c4eeb526 | testsuite/test_util.py | testsuite/test_util.py | import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*'])
| import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
self.assertEquals(pep8.normalize_paths([]), [])
self.assertEquals(pep8.normalize_paths(None), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*'])
| Add a few more cases of "not value" | Add a few more cases of "not value"
| Python | mit | ojengwa/pep8,pedros/pep8,asandyz/pep8,jayvdb/pep8,doismellburning/pep8,pandeesh/pep8,jayvdb/pep8,PyCQA/pep8,ABaldwinHunter/pep8,codeclimate/pep8,ABaldwinHunter/pep8-clone-classic,zevnux/pep8,MeteorAdminz/pep8 | import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
+ self.assertEquals(pep8.normalize_paths([]), [])
+ self.assertEquals(pep8.normalize_paths(None), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
+ self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*'])
| Add a few more cases of "not value" | ## Code Before:
import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*'])
## Instruction:
Add a few more cases of "not value"
## Code After:
import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
self.assertEquals(pep8.normalize_paths([]), [])
self.assertEquals(pep8.normalize_paths(None), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*'])
| import os
import unittest
import pep8
class UtilTestCase(unittest.TestCase):
def test_normalize_paths(self):
cwd = os.getcwd()
self.assertEquals(pep8.normalize_paths(''), [])
+ self.assertEquals(pep8.normalize_paths([]), [])
+ self.assertEquals(pep8.normalize_paths(None), [])
self.assertEquals(pep8.normalize_paths(['foo']), ['foo'])
self.assertEquals(pep8.normalize_paths('foo'), ['foo'])
self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar'])
+ self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar'])
self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'),
['/foo/bar', cwd + '/bat'])
self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"),
['.pyc', cwd + '/build/*']) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.