commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
4d19cc36e866c8e21a526cd228f170ffd177292b
run_ctest.py
run_ctest.py
import platform import os import subprocess import sys if platform.system() == "Windows": import distutils.msvc9compiler as msvc if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if len(sys.argv) > 1: msvc.find_vcvarsall = lambda _: sys.argv[1] envs = msvc.query_vcvarsall(sys.argv[2]) for k,v in envs.items(): k = k.upper() v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";")) v = v.replace("'\''",r"'\'\\\'\''") print "export %(k)s='\''%(v)s'\''" % locals() elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.")
import platform import os import subprocess if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.")
Test travisci windows build workaround. (3)
Test travisci windows build workaround. (3)
Python
unlicense
h-s-c/ci-tools
import platform import os import subprocess - import sys - - if platform.system() == "Windows": - import distutils.msvc9compiler as msvc if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) - if len(sys.argv) > 1: - msvc.find_vcvarsall = lambda _: sys.argv[1] - envs = msvc.query_vcvarsall(sys.argv[2]) - for k,v in envs.items(): - k = k.upper() - v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";")) - v = v.replace("'\''",r"'\'\\\'\''") - print "export %(k)s='\''%(v)s'\''" % locals() elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.")
Test travisci windows build workaround. (3)
## Code Before: import platform import os import subprocess import sys if platform.system() == "Windows": import distutils.msvc9compiler as msvc if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if len(sys.argv) > 1: msvc.find_vcvarsall = lambda _: sys.argv[1] envs = msvc.query_vcvarsall(sys.argv[2]) for k,v in envs.items(): k = k.upper() v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";")) v = v.replace("'\''",r"'\'\\\'\''") print "export %(k)s='\''%(v)s'\''" % locals() elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.") ## Instruction: Test travisci windows build workaround. (3) ## Code After: import platform import os import subprocess if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.")
import platform import os import subprocess - import sys - - if platform.system() == "Windows": - import distutils.msvc9compiler as msvc if __name__ == "__main__": CITOOLS_PATH = os.path.join(os.getcwd(), "ci-tools") CMAKE_PATH = os.path.join(CITOOLS_PATH, "cmake") if platform.system() == "Linux": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) elif platform.system() == "Windows": os.environ["PATH"] = os.path.join(CMAKE_PATH, "bin")+";"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) - if len(sys.argv) > 1: - msvc.find_vcvarsall = lambda _: sys.argv[1] - envs = msvc.query_vcvarsall(sys.argv[2]) - for k,v in envs.items(): - k = k.upper() - v = ":".join(subprocess.check_output(["cygpath","-u",p]).rstrip() for p in v.split(";")) - v = v.replace("'\''",r"'\'\\\'\''") - print "export %(k)s='\''%(v)s'\''" % locals() elif platform.system() == "Darwin": os.environ["PATH"] = os.path.join(CMAKE_PATH, "CMake.app", "Contents", "bin")+":"+os.environ.get("PATH", os.path.join(CMAKE_PATH, "bin")) if subprocess.call("ctest -VV -S ci-tools/run_ctest.cmake", shell=True) != 0: raise Exception("CTest returned an error.")
b1bf5dfa91f1f7b84512f72d6e5e18c2109f3239
addic7ed/__init__.py
addic7ed/__init__.py
from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": print(ep.rename(filename), end="\n\n") else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % filename, "green")) except Exception as e: print(colored(e, "red"), end="\n\n")
from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": print(ep.rename(filename)) else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % filename, "green"), end="\n\n") except Exception as e: print(colored(e, "red"), end="\n\n")
Fix newline output of downloaded srt
Fix newline output of downloaded srt
Python
mit
Jesus-21/addic7ed
from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": - print(ep.rename(filename), end="\n\n") + print(ep.rename(filename)) else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % - filename, "green")) + filename, "green"), end="\n\n") except Exception as e: print(colored(e, "red"), end="\n\n")
Fix newline output of downloaded srt
## Code Before: from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": print(ep.rename(filename), end="\n\n") else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % filename, "green")) except Exception as e: print(colored(e, "red"), end="\n\n") ## Instruction: Fix newline output of downloaded srt ## Code After: from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": print(ep.rename(filename)) else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % filename, "green"), end="\n\n") except Exception as e: print(colored(e, "red"), end="\n\n")
from termcolor import colored from .parser import Addic7edParser from .file_crawler import FileCrawler from .logger import init_logger from .config import Config def addic7ed(): try: init_logger() Config.load() main() except (EOFError, KeyboardInterrupt, SystemExit): print(colored("\nBye!", "yellow")) exit(0) def main(): crawler = FileCrawler() parser = Addic7edParser() for filename, ep in crawler.episodes.items(): subs = parser.parse(**ep.infos) print(ep) if not subs: print(colored("No subtitles for %s" % filename, "red"), end="\n\n") continue for i, sub in enumerate(subs): print("[%s] %s" % (colored(i, "yellow"), sub)) if Config.dry_run: print() continue else: version = input('Download number? ') if not version: print(colored("Nothing to do!", "yellow"), end="\n\n") continue try: if Config.rename != "sub": filename = subs[int(version)].download() if filename and Config.rename == "video": - print(ep.rename(filename), end="\n\n") ? ------------ + print(ep.rename(filename)) else: filename = subs[int(version)].download("%s.srt" % ep.filename) print(colored("Downloaded %s subtitle file" % - filename, "green")) + filename, "green"), end="\n\n") ? ++++++++++++ except Exception as e: print(colored(e, "red"), end="\n\n")
e790e47e6b87bc2e49e8b74d491eb023c4468254
src/sentry/web/frontend/csrf_failure.py
src/sentry/web/frontend/csrf_failure.py
from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER from sentry.web.frontend.base import BaseView class CsrfFailureView(BaseView): auth_required = False sudo_required = False def handle(self, request, reason=""): context = { 'no_referer': reason == REASON_NO_REFERER } return self.respond('sentry/403-csrf-failure.html', status=403) view = CsrfFailureView.as_view()
from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.utils.decorators import method_decorator from sentry.web.helpers import render_to_response class CsrfFailureView(View): @method_decorator(csrf_exempt) def dispatch(self, request, reason=""): context = { 'no_referer': reason == REASON_NO_REFERER, 'request': request, } return render_to_response('sentry/403-csrf-failure.html', context, request, status=403) view = CsrfFailureView.as_view()
Kill possible recursion on csrf decorator
Kill possible recursion on csrf decorator
Python
bsd-3-clause
boneyao/sentry,jean/sentry,boneyao/sentry,mvaled/sentry,felixbuenemann/sentry,kevinlondon/sentry,TedaLIEz/sentry,JamesMura/sentry,kevinastone/sentry,korealerts1/sentry,JackDanger/sentry,songyi199111/sentry,songyi199111/sentry,fuziontech/sentry,JamesMura/sentry,BuildingLink/sentry,camilonova/sentry,wujuguang/sentry,argonemyth/sentry,wujuguang/sentry,pauloschilling/sentry,zenefits/sentry,nicholasserra/sentry,beeftornado/sentry,ewdurbin/sentry,gg7/sentry,Natim/sentry,vperron/sentry,Natim/sentry,korealerts1/sentry,kevinlondon/sentry,alexm92/sentry,wong2/sentry,gencer/sentry,BayanGroup/sentry,fuziontech/sentry,jean/sentry,JTCunning/sentry,alexm92/sentry,drcapulet/sentry,gencer/sentry,ifduyue/sentry,ewdurbin/sentry,imankulov/sentry,felixbuenemann/sentry,hongliang5623/sentry,wujuguang/sentry,pauloschilling/sentry,drcapulet/sentry,looker/sentry,nicholasserra/sentry,Kryz/sentry,mvaled/sentry,ewdurbin/sentry,wong2/sentry,imankulov/sentry,jean/sentry,kevinastone/sentry,1tush/sentry,mvaled/sentry,fotinakis/sentry,1tush/sentry,gencer/sentry,hongliang5623/sentry,vperron/sentry,looker/sentry,JackDanger/sentry,hongliang5623/sentry,zenefits/sentry,nicholasserra/sentry,zenefits/sentry,jokey2k/sentry,JamesMura/sentry,songyi199111/sentry,1tush/sentry,drcapulet/sentry,ngonzalvez/sentry,mvaled/sentry,Kryz/sentry,BayanGroup/sentry,kevinlondon/sentry,daevaorn/sentry,ifduyue/sentry,ifduyue/sentry,fotinakis/sentry,zenefits/sentry,BuildingLink/sentry,camilonova/sentry,JTCunning/sentry,TedaLIEz/sentry,Natim/sentry,ngonzalvez/sentry,llonchj/sentry,BuildingLink/sentry,argonemyth/sentry,jean/sentry,argonemyth/sentry,wong2/sentry,looker/sentry,looker/sentry,vperron/sentry,TedaLIEz/sentry,ngonzalvez/sentry,JamesMura/sentry,camilonova/sentry,beeftornado/sentry,looker/sentry,jokey2k/sentry,zenefits/sentry,BayanGroup/sentry,JackDanger/sentry,daevaorn/sentry,BuildingLink/sentry,gg7/sentry,BuildingLink/sentry,Kryz/sentry,gg7/sentry,JTCunning/sentry,boneyao/sentry,mitsuhiko/sentry,ifduyue/sentry,llonchj/sentry,mvaled/sentry,fotinakis/sentry,korealerts1/sentry,jean/sentry,mvaled/sentry,llonchj/sentry,mitsuhiko/sentry,felixbuenemann/sentry,beeftornado/sentry,fuziontech/sentry,JamesMura/sentry,kevinastone/sentry,imankulov/sentry,daevaorn/sentry,jokey2k/sentry,gencer/sentry,pauloschilling/sentry,gencer/sentry,ifduyue/sentry,fotinakis/sentry,alexm92/sentry,daevaorn/sentry
from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER + from django.views.decorators.csrf import csrf_exempt + from django.views.generic import View + from django.utils.decorators import method_decorator - from sentry.web.frontend.base import BaseView + from sentry.web.helpers import render_to_response - class CsrfFailureView(BaseView): + class CsrfFailureView(View): + @method_decorator(csrf_exempt) - auth_required = False - sudo_required = False - - def handle(self, request, reason=""): + def dispatch(self, request, reason=""): context = { - 'no_referer': reason == REASON_NO_REFERER + 'no_referer': reason == REASON_NO_REFERER, + 'request': request, } - return self.respond('sentry/403-csrf-failure.html', status=403) + return render_to_response('sentry/403-csrf-failure.html', context, request, + status=403) view = CsrfFailureView.as_view()
Kill possible recursion on csrf decorator
## Code Before: from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER from sentry.web.frontend.base import BaseView class CsrfFailureView(BaseView): auth_required = False sudo_required = False def handle(self, request, reason=""): context = { 'no_referer': reason == REASON_NO_REFERER } return self.respond('sentry/403-csrf-failure.html', status=403) view = CsrfFailureView.as_view() ## Instruction: Kill possible recursion on csrf decorator ## Code After: from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER from django.views.decorators.csrf import csrf_exempt from django.views.generic import View from django.utils.decorators import method_decorator from sentry.web.helpers import render_to_response class CsrfFailureView(View): @method_decorator(csrf_exempt) def dispatch(self, request, reason=""): context = { 'no_referer': reason == REASON_NO_REFERER, 'request': request, } return render_to_response('sentry/403-csrf-failure.html', context, request, status=403) view = CsrfFailureView.as_view()
from __future__ import absolute_import from django.middleware.csrf import REASON_NO_REFERER + from django.views.decorators.csrf import csrf_exempt + from django.views.generic import View + from django.utils.decorators import method_decorator - from sentry.web.frontend.base import BaseView + from sentry.web.helpers import render_to_response - class CsrfFailureView(BaseView): ? ---- + class CsrfFailureView(View): + @method_decorator(csrf_exempt) - auth_required = False - sudo_required = False - - def handle(self, request, reason=""): ? ----- + def dispatch(self, request, reason=""): ? +++++++ context = { - 'no_referer': reason == REASON_NO_REFERER + 'no_referer': reason == REASON_NO_REFERER, ? + + 'request': request, } - return self.respond('sentry/403-csrf-failure.html', status=403) + return render_to_response('sentry/403-csrf-failure.html', context, request, + status=403) view = CsrfFailureView.as_view()
33f4036825c6ff4d9df0038471727648e0df100d
feder/virus_scan/engine/base.py
feder/virus_scan/engine/base.py
from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), self.signer.sign(self.name), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
import urllib.parse from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), urllib.parse.quote(self.signer.sign(self.name)), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
Fix urlencode in webhook url
Fix urlencode in webhook url
Python
mit
watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder
+ import urllib.parse + from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), - self.signer.sign(self.name), + urllib.parse.quote(self.signer.sign(self.name)), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
Fix urlencode in webhook url
## Code Before: from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), self.signer.sign(self.name), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) ) ## Instruction: Fix urlencode in webhook url ## Code After: import urllib.parse from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), urllib.parse.quote(self.signer.sign(self.name)), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
+ import urllib.parse + from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), - self.signer.sign(self.name), + urllib.parse.quote(self.signer.sign(self.name)), ? +++++++++++++++++++ + ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
c621bc7c94dbbeb5540b2ce46437ee24ecbc33dd
test/test_interface.py
test/test_interface.py
from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service")
import cloudbridge from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service") def test_library_version(self): """ Check that the library version can be retrieved. """ self.assertIsNotNone(cloudbridge.get_version(), "Did not get library version.")
Add a library version test
Add a library version test
Python
mit
gvlproject/cloudbridge,ms-azure-cloudbroker/cloudbridge,gvlproject/libcloudbridge
+ import cloudbridge from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service") + def test_library_version(self): + """ + Check that the library version can be retrieved. + """ + self.assertIsNotNone(cloudbridge.get_version(), + "Did not get library version.") +
Add a library version test
## Code Before: from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service") ## Instruction: Add a library version test ## Code After: import cloudbridge from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service") def test_library_version(self): """ Check that the library version can be retrieved. """ self.assertIsNotNone(cloudbridge.get_version(), "Did not get library version.")
+ import cloudbridge from cloudbridge.cloud import interfaces from test.helpers import ProviderTestBase class CloudInterfaceTestCase(ProviderTestBase): def __init__(self, methodName, provider): super(CloudInterfaceTestCase, self).__init__( methodName=methodName, provider=provider) def test_name_property(self): """ Name should always return a value and should not raise an exception """ assert self.provider.name def test_has_service_valid_service_type(self): """ has_service with a valid service type should return a boolean and raise no exceptions """ for key, value in interfaces.CloudServiceType.__dict__.items(): if not key.startswith("__"): self.provider.has_service(value) def test_has_service_invalid_service_type(self): """ has_service with an invalid service type should return False """ self.assertFalse( self.provider.has_service("NON_EXISTENT_SERVICE"), "has_service should not return True for a non-existent service") + + def test_library_version(self): + """ + Check that the library version can be retrieved. + """ + self.assertIsNotNone(cloudbridge.get_version(), + "Did not get library version.")
50f8efd7bcbf032fd0295c460b98640d0bf6c1ed
smithers/smithers/conf/server.py
smithers/smithers/conf/server.py
from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' STATSD_HOST = 'graphite1.private.phx1.mozilla.com' STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500
from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com') STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500
Make statsd host configurable via env.
Make statsd host configurable via env.
Python
mpl-2.0
mozilla/mrburns,mozilla/mrburns,mozilla/mrburns
from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' - STATSD_HOST = 'graphite1.private.phx1.mozilla.com' + STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com') STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500
Make statsd host configurable via env.
## Code Before: from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' STATSD_HOST = 'graphite1.private.phx1.mozilla.com' STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500 ## Instruction: Make statsd host configurable via env. ## Code After: from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com') STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500
from os import getenv GEOIP_DB_FILE = '/usr/local/share/GeoIP/GeoIP2-City.mmdb' - STATSD_HOST = 'graphite1.private.phx1.mozilla.com' + STATSD_HOST = getenv('STATSD_HOST', 'graphite1.private.phx1.mozilla.com') ? ++++++++++++++++++++++ + STATSD_PORT = 8125 STATSD_PREFIX = 'glow-workers-{}'.format(getenv('DJANGO_SERVER_ENV')) COUNTRY_MIN_SHARE = 500
1b95969110f97af397cb3314b59c30679911da48
scripts/scrape-cdc-state-case-counts.py
scripts/scrape-cdc-state-case-counts.py
import requests import lxml.html import pandas as pd import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") cells = [ [ td.text_content().strip() for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
import requests import lxml.html import pandas as pd import re import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS paren_pat = re.compile(r"\([^\)]+\)") def parse_cell(text): return re.sub(paren_pat, "", text).strip() def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") cells = [ [ parse_cell(td.text_content()) for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
Update CDC scraper to handle new format
Update CDC scraper to handle new format
Python
mit
BuzzFeedNews/zika-data
import requests import lxml.html import pandas as pd + import re import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS + + paren_pat = re.compile(r"\([^\)]+\)") + + def parse_cell(text): + return re.sub(paren_pat, "", text).strip() def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") - cells = [ [ td.text_content().strip() + cells = [ [ parse_cell(td.text_content()) for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
Update CDC scraper to handle new format
## Code Before: import requests import lxml.html import pandas as pd import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") cells = [ [ td.text_content().strip() for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8") ## Instruction: Update CDC scraper to handle new format ## Code After: import requests import lxml.html import pandas as pd import re import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS paren_pat = re.compile(r"\([^\)]+\)") def parse_cell(text): return re.sub(paren_pat, "", text).strip() def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") cells = [ [ parse_cell(td.text_content()) for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
import requests import lxml.html import pandas as pd + import re import sys URL = "http://www.cdc.gov/zika/geo/united-states.html" INT_COLS = [ "travel_associated_cases", "locally_acquired_cases" ] COLS = [ "state_or_territory" ] + INT_COLS + + paren_pat = re.compile(r"\([^\)]+\)") + + def parse_cell(text): + return re.sub(paren_pat, "", text).strip() def scrape(): html = requests.get(URL).content dom = lxml.html.fromstring(html) table = dom.cssselect("table")[0] rows = table.cssselect("tr") - cells = [ [ td.text_content().strip() ? ------- - + cells = [ [ parse_cell(td.text_content()) ? +++++++++++ for td in tr.cssselect("td") ] for tr in rows ] data = [ c for c in cells if sum(len(x) != 0 for x in c) == 3 ] df = pd.DataFrame(data, columns=COLS) df[INT_COLS] = df[INT_COLS].astype(int) return df if __name__ == "__main__": df = scrape() df.to_csv(sys.stdout, index=False, encoding="utf-8")
524d5427d54342f26008a5b527140d4158f70edf
tests/test_extension.py
tests/test_extension.py
from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info
from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() get_websocket().data = None # make sure it's cleared frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info
Clear websocket data to try and fix Travis
Clear websocket data to try and fix Travis
Python
agpl-3.0
palfrey/mopidy-tachikoma,palfrey/mopidy-tachikoma
from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() + get_websocket().data = None # make sure it's cleared frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info
Clear websocket data to try and fix Travis
## Code Before: from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info ## Instruction: Clear websocket data to try and fix Travis ## Code After: from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() get_websocket().data = None # make sure it's cleared frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info
from __future__ import unicode_literals import json from test_helpers import MockTrack, get_websocket, make_frontend, patched_bot from mopidy_tachikoma import Extension def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[tachikoma]' in config assert 'enabled = true' in config assert 'slack_token = ' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'slack_token' in schema @patched_bot def test_can_connect(): make_frontend() @patched_bot def test_gets_events(): frontend = make_frontend() frontend.doSlackLoop( None, MockTrack(), [{"type": "message", "channel": "mock_channel"}]) data = json.loads(get_websocket().data) assert { 'channel': 'mock_channel', 'text': 'Now playing *foo* from *bar*', 'type': 'message'} == data @patched_bot def test_says_one_thing_per_channel(): frontend = make_frontend() song = MockTrack() + get_websocket().data = None # make sure it's cleared frontend.doSlackLoop( song, song, [{"type": "message", "channel": "mock_channel"}]) assert get_websocket().data is None # same song, no info
64d7ca9695eed6112c793fda3f2e7fea3751c3cc
tasks.py
tasks.py
from invoke import run from invoke import task @task def clean(docs=False, bytecode=True, extra=''): patterns = ['build'] if docs: patterns.append('docs/_build') if bytecode: patterns.append('**/*.pyc') if extra: patterns.append(extra) for pattern in patterns: run("rm -rf %s" % pattern) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8")
from invoke import run from invoke import task @task def clean(all=False): if all: flag = "--all" else: flag = "" run("python setup.py clean {}".format(flag)) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8")
Change clean task to use setup.py
Change clean task to use setup.py
Python
bsd-3-clause
pando85/django-registration,allo-/django-registration,sergafts/django-registration,pando85/django-registration,allo-/django-registration,sergafts/django-registration
from invoke import run from invoke import task @task + def clean(all=False): - def clean(docs=False, bytecode=True, extra=''): - patterns = ['build'] - if docs: - patterns.append('docs/_build') - if bytecode: - patterns.append('**/*.pyc') - if extra: + if all: - patterns.append(extra) - for pattern in patterns: - run("rm -rf %s" % pattern) + flag = "--all" + else: + flag = "" + run("python setup.py clean {}".format(flag)) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8")
Change clean task to use setup.py
## Code Before: from invoke import run from invoke import task @task def clean(docs=False, bytecode=True, extra=''): patterns = ['build'] if docs: patterns.append('docs/_build') if bytecode: patterns.append('**/*.pyc') if extra: patterns.append(extra) for pattern in patterns: run("rm -rf %s" % pattern) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8") ## Instruction: Change clean task to use setup.py ## Code After: from invoke import run from invoke import task @task def clean(all=False): if all: flag = "--all" else: flag = "" run("python setup.py clean {}".format(flag)) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8")
from invoke import run from invoke import task @task + def clean(all=False): - def clean(docs=False, bytecode=True, extra=''): - patterns = ['build'] - if docs: - patterns.append('docs/_build') - if bytecode: - patterns.append('**/*.pyc') - if extra: ? ---- + if all: ? ++ - patterns.append(extra) - for pattern in patterns: - run("rm -rf %s" % pattern) + flag = "--all" + else: + flag = "" + run("python setup.py clean {}".format(flag)) @task def build(docs=False): run("python setup.py build") if docs: run("sphinx-build docs docs/_build") @task def test(): run("python setup.py test") @task def lint(): run("flake8")
f794c6ed1f6be231d79ac35759ad76270c3e14e0
brains/mapping/admin.py
brains/mapping/admin.py
from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', ('reported_by', 'origin', 'reported_date') )} ),) readonly_fields = ['players', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin)
from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', ('reported_by', 'origin'), 'reported_date', )} ),) readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined', 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level', 'players', 'reported_by', 'origin', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin)
Set everything on the report read only.
Set everything on the report read only.
Python
bsd-3-clause
crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains,crisisking/udbraaains
from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', - ('reported_by', 'origin', 'reported_date') + ('reported_by', 'origin'), + 'reported_date', )} ),) - readonly_fields = ['players', 'reported_date'] + + readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined', + 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level', + 'players', 'reported_by', 'origin', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin)
Set everything on the report read only.
## Code Before: from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', ('reported_by', 'origin', 'reported_date') )} ),) readonly_fields = ['players', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin) ## Instruction: Set everything on the report read only. ## Code After: from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', ('reported_by', 'origin'), 'reported_date', )} ),) readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined', 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level', 'players', 'reported_by', 'origin', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin)
from django.contrib import admin from mapping.models import Location, Report class LocationAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ( ('name', 'suburb'), ('x', 'y'), 'building_type' )} ),) list_display = ['name', 'x', 'y', 'suburb'] list_filter = ['suburb'] search_fields = ['name'] readonly_fields = ['x', 'y', 'name', 'building_type', 'suburb'] actions = None def has_add_permission(self, request): return False class ReportAdmin(admin.ModelAdmin): fieldsets = ((None, {'fields': ('location', ('zombies_only', 'inside'), ('is_ruined', 'is_illuminated', 'has_tree'), ('zombies_present', 'barricade_level'), 'players', - ('reported_by', 'origin', 'reported_date') ? ----------------- + ('reported_by', 'origin'), ? + + 'reported_date', )} ),) - readonly_fields = ['players', 'reported_date'] + + readonly_fields = ['location', 'zombies_only', 'inside', 'is_ruined', + 'is_illuminated', 'has_tree', 'zombies_present', 'barricade_level', + 'players', 'reported_by', 'origin', 'reported_date'] admin.site.register(Location, LocationAdmin) admin.site.register(Report, ReportAdmin)
7a97694d14e9ed02ae38ab713267d4c722079c9c
kolibri/core/webpack/test/test_webpack_tags.py
kolibri/core/webpack/test/test_webpack_tags.py
from __future__ import absolute_import, print_function, unicode_literals from django.test import TestCase from ..hooks import WebpackBundleHook class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" entry_file = "assets/src/kolibri_core_app.js" class KolibriTagNavigationTestCase(TestCase): def test_frontend_tag(self): self.assertIn( "non_default_frontend", TestHook().render_to_html() )
from __future__ import absolute_import, print_function, unicode_literals import json import tempfile from django.test import TestCase from ..hooks import WebpackBundleHook TEST_STATS_FILE = None class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" src_file = "assets/src/kolibri_core_app.js" @property def stats_file(self): return TEST_STATS_FILE.name class KolibriTagNavigationTestCase(TestCase): def setUp(self): global TEST_STATS_FILE TestCase.setUp(self) TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False) self.test_hook = TestHook() json.dump( { "status": "done", "chunks": { "non_default_frontend": [ { "name": "non_default_frontend-2c4fb3d6a29238b06f84.js", "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js", "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js" } ] }, "publicPath": "default_frontend/" }, TEST_STATS_FILE ) TEST_STATS_FILE.close() def test_frontend_tag(self): self.assertIn( "non_default_frontend", self.test_hook.render_to_html() )
Create a faked hook that does not try to read a live stats file
Create a faked hook that does not try to read a live stats file
Python
mit
jonboiser/kolibri,jtamiace/kolibri,66eli77/kolibri,66eli77/kolibri,mrpau/kolibri,rtibbles/kolibri,learningequality/kolibri,lyw07/kolibri,jtamiace/kolibri,jtamiace/kolibri,mrpau/kolibri,ralphiee22/kolibri,jayoshih/kolibri,learningequality/kolibri,DXCanas/kolibri,jonboiser/kolibri,benjaoming/kolibri,christianmemije/kolibri,indirectlylit/kolibri,benjaoming/kolibri,lyw07/kolibri,aronasorman/kolibri,66eli77/kolibri,MCGallaspy/kolibri,jayoshih/kolibri,DXCanas/kolibri,jonboiser/kolibri,MCGallaspy/kolibri,aronasorman/kolibri,whitzhu/kolibri,DXCanas/kolibri,whitzhu/kolibri,benjaoming/kolibri,rtibbles/kolibri,mrpau/kolibri,lyw07/kolibri,learningequality/kolibri,jamalex/kolibri,rtibbles/kolibri,lyw07/kolibri,MingDai/kolibri,rtibbles/kolibri,jonboiser/kolibri,indirectlylit/kolibri,jamalex/kolibri,whitzhu/kolibri,MingDai/kolibri,66eli77/kolibri,MingDai/kolibri,indirectlylit/kolibri,MingDai/kolibri,jamalex/kolibri,DXCanas/kolibri,jtamiace/kolibri,jamalex/kolibri,MCGallaspy/kolibri,christianmemije/kolibri,ralphiee22/kolibri,aronasorman/kolibri,aronasorman/kolibri,whitzhu/kolibri,jayoshih/kolibri,ralphiee22/kolibri,indirectlylit/kolibri,christianmemije/kolibri,learningequality/kolibri,benjaoming/kolibri,mrpau/kolibri,christianmemije/kolibri,ralphiee22/kolibri,jayoshih/kolibri
from __future__ import absolute_import, print_function, unicode_literals + + import json + import tempfile + from django.test import TestCase from ..hooks import WebpackBundleHook + TEST_STATS_FILE = None + + class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" - entry_file = "assets/src/kolibri_core_app.js" + src_file = "assets/src/kolibri_core_app.js" + + @property + def stats_file(self): + return TEST_STATS_FILE.name class KolibriTagNavigationTestCase(TestCase): + def setUp(self): + global TEST_STATS_FILE + TestCase.setUp(self) + TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False) + self.test_hook = TestHook() + json.dump( + { + "status": "done", + "chunks": { + "non_default_frontend": [ + { + "name": "non_default_frontend-2c4fb3d6a29238b06f84.js", + "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js", + "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js" + } + ] + }, + "publicPath": "default_frontend/" + }, + TEST_STATS_FILE + ) + TEST_STATS_FILE.close() + def test_frontend_tag(self): self.assertIn( "non_default_frontend", - TestHook().render_to_html() + self.test_hook.render_to_html() )
Create a faked hook that does not try to read a live stats file
## Code Before: from __future__ import absolute_import, print_function, unicode_literals from django.test import TestCase from ..hooks import WebpackBundleHook class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" entry_file = "assets/src/kolibri_core_app.js" class KolibriTagNavigationTestCase(TestCase): def test_frontend_tag(self): self.assertIn( "non_default_frontend", TestHook().render_to_html() ) ## Instruction: Create a faked hook that does not try to read a live stats file ## Code After: from __future__ import absolute_import, print_function, unicode_literals import json import tempfile from django.test import TestCase from ..hooks import WebpackBundleHook TEST_STATS_FILE = None class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" src_file = "assets/src/kolibri_core_app.js" @property def stats_file(self): return TEST_STATS_FILE.name class KolibriTagNavigationTestCase(TestCase): def setUp(self): global TEST_STATS_FILE TestCase.setUp(self) TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False) self.test_hook = TestHook() json.dump( { "status": "done", "chunks": { "non_default_frontend": [ { "name": "non_default_frontend-2c4fb3d6a29238b06f84.js", "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js", "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js" } ] }, "publicPath": "default_frontend/" }, TEST_STATS_FILE ) TEST_STATS_FILE.close() def test_frontend_tag(self): self.assertIn( "non_default_frontend", self.test_hook.render_to_html() )
from __future__ import absolute_import, print_function, unicode_literals + + import json + import tempfile + from django.test import TestCase from ..hooks import WebpackBundleHook + TEST_STATS_FILE = None + + class TestHook(WebpackBundleHook): unique_slug = "non_default_frontend" - entry_file = "assets/src/kolibri_core_app.js" ? ^^^ ^ + src_file = "assets/src/kolibri_core_app.js" ? ^ ^ + + @property + def stats_file(self): + return TEST_STATS_FILE.name class KolibriTagNavigationTestCase(TestCase): + def setUp(self): + global TEST_STATS_FILE + TestCase.setUp(self) + TEST_STATS_FILE = tempfile.NamedTemporaryFile(mode='w+', delete=False) + self.test_hook = TestHook() + json.dump( + { + "status": "done", + "chunks": { + "non_default_frontend": [ + { + "name": "non_default_frontend-2c4fb3d6a29238b06f84.js", + "publicPath": "non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js", + "path": "kolibri/core/static/non_default_frontend/non_default_frontend-2c4fb3d6a29238b06f84.js" + } + ] + }, + "publicPath": "default_frontend/" + }, + TEST_STATS_FILE + ) + TEST_STATS_FILE.close() + def test_frontend_tag(self): self.assertIn( "non_default_frontend", - TestHook().render_to_html() ? ^ ^ -- + self.test_hook.render_to_html() ? ^^^^^^ ^^ )
91f107ef2ebdaf7ff210b9f36e2c810441f389e7
services/rdio.py
services/rdio.py
from werkzeug.urls import url_decode from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False signature_type = SIGNATURE_TYPE_BODY def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key'])
from werkzeug.urls import url_decode import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key'])
Allow Rdio to use default signature handling
Allow Rdio to use default signature handling
Python
bsd-3-clause
foauth/oauth-proxy,foauth/foauth.org,foauth/foauth.org,foauth/foauth.org
from werkzeug.urls import url_decode - from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False - signature_type = SIGNATURE_TYPE_BODY def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key'])
Allow Rdio to use default signature handling
## Code Before: from werkzeug.urls import url_decode from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False signature_type = SIGNATURE_TYPE_BODY def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key']) ## Instruction: Allow Rdio to use default signature handling ## Code After: from werkzeug.urls import url_decode import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key'])
from werkzeug.urls import url_decode - from oauthlib.oauth1.rfc5849 import SIGNATURE_TYPE_BODY import foauth.providers class Rdio(foauth.providers.OAuth1): # General info about the provider provider_url = 'http://www.rdio.com/' docs_url = 'http://developer.rdio.com/docs/REST/' category = 'Music' # URLs to interact with the API request_token_url = 'http://api.rdio.com/oauth/request_token' authorize_url = None # Provided when the request token is granted access_token_url = 'http://api.rdio.com/oauth/access_token' api_domain = 'api.rdio.com' available_permissions = [ (None, 'access and manage your music'), ] https = False - signature_type = SIGNATURE_TYPE_BODY def parse_token(self, content): # Override standard token request to also get the authorization URL data = url_decode(content) if 'login_url' in data: self.authorize_url = data['login_url'] return super(Rdio, self).parse_token(content) def get_user_id(self, key): r = self.api(key, self.api_domain, u'/1/', method='POST', data={ 'method': 'currentUser', }) return unicode(r.json[u'result'][u'key'])
aeaf802100cd6869178dd9f412d35e452916a63d
common/commands/view_manipulation.py
common/commands/view_manipulation.py
from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) self.view.run_command("_enter_insert_mode") class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True)
from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) # NeoVintageous renamed the command starting with v1.22.0. # We call both commands for backwards compatibility. self.view.run_command("_enter_insert_mode") self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0 class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True)
Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0
Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0 Fixes #1395 In NeoVintageous/NeoVintageous#749, pushed as 1.22.0 (Oct 2020), the relevant commands were renamed. We follow the new names, but for now also call the old ones.
Python
mit
divmain/GitSavvy,divmain/GitSavvy,divmain/GitSavvy
from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) + # NeoVintageous renamed the command starting with v1.22.0. + # We call both commands for backwards compatibility. self.view.run_command("_enter_insert_mode") + self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0 class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True)
Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0
## Code Before: from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) self.view.run_command("_enter_insert_mode") class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True) ## Instruction: Fix `vintageous_enter_insert_mode` for NeoVintageous 1.22.0 ## Code After: from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) # NeoVintageous renamed the command starting with v1.22.0. # We call both commands for backwards compatibility. self.view.run_command("_enter_insert_mode") self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0 class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True)
from sublime_plugin import TextCommand from ...core.git_command import GitCommand __all__ = ( "gs_handle_vintageous", "gs_handle_arrow_keys" ) class gs_handle_vintageous(TextCommand, GitCommand): """ Set the vintageous_friendly view setting if needed. Enter insert mode if vintageous_enter_insert_mode option is enabled. """ def run(self, edit): if self.savvy_settings.get("vintageous_friendly"): self.view.settings().set("git_savvy.vintageous_friendly", True) if self.savvy_settings.get("vintageous_enter_insert_mode"): self.view.settings().set("vintageous_reset_mode_when_switching_tabs", False) + # NeoVintageous renamed the command starting with v1.22.0. + # We call both commands for backwards compatibility. self.view.run_command("_enter_insert_mode") + self.view.run_command("nv_enter_insert_mode") # since NeoVintageous 1.22.0 class gs_handle_arrow_keys(TextCommand, GitCommand): """ Set the arrow_keys_navigation view setting if needed. It allows navigation by using arrow keys. """ def run(self, edit): if self.savvy_settings.get("arrow_keys_navigation"): self.view.settings().set("git_savvy.arrow_keys_navigation", True)
4d414fe592bfd7f085f9aaea0b6992d28ad193ce
tcconfig/_common.py
tcconfig/_common.py
from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( "network interface not found: " + device) def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network
from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( "network interface not found: {}".format(device)) def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network
Change to use format method
Change to use format method
Python
mit
thombashi/tcconfig,thombashi/tcconfig
from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( - "network interface not found: " + device) + "network interface not found: {}".format(device)) def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network
Change to use format method
## Code Before: from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( "network interface not found: " + device) def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network ## Instruction: Change to use format method ## Code After: from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( "network interface not found: {}".format(device)) def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network
from __future__ import absolute_import import dataproperty import six from ._error import NetworkInterfaceNotFoundError ANYWHERE_NETWORK = "0.0.0.0/0" def verify_network_interface(device): try: import netifaces except ImportError: return if device not in netifaces.interfaces(): raise NetworkInterfaceNotFoundError( - "network interface not found: " + device) ? ^^^ + "network interface not found: {}".format(device)) ? ++ ^^^^^^^^ + def sanitize_network(network): """ :return: Network string :rtype: str :raises ValueError: if the network string is invalid. """ import ipaddress if dataproperty.is_empty_string(network): return "" if network.lower() == "anywhere": return ANYWHERE_NETWORK try: ipaddress.IPv4Address(six.u(network)) return network + "/32" except ipaddress.AddressValueError: pass ipaddress.IPv4Network(six.u(network)) # validate network str return network
5a03cd340e5dc8a796c7d430128f0e22be17333e
qiime/sdk/__init__.py
qiime/sdk/__init__.py
from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type']
from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type'] # Various URLs CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131' HELP_URL = 'http://2.qiime.org' CONDA_CHANNEL = 'https://anaconda.org/qiime2'
Add helper URLs to qiime.sdk
ENH: Add helper URLs to qiime.sdk Adds citation url, help page, and conda channel URLs to qiime.sdk
Python
bsd-3-clause
biocore/qiime2,thermokarst/qiime2,ebolyen/qiime2,jakereps/qiime2,qiime2/qiime2,qiime2/qiime2,nervous-laughter/qiime2,biocore/qiime2,thermokarst/qiime2,jairideout/qiime2,jakereps/qiime2
from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type'] + # Various URLs + CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131' + HELP_URL = 'http://2.qiime.org' + CONDA_CHANNEL = 'https://anaconda.org/qiime2' +
Add helper URLs to qiime.sdk
## Code Before: from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type'] ## Instruction: Add helper URLs to qiime.sdk ## Code After: from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type'] # Various URLs CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131' HELP_URL = 'http://2.qiime.org' CONDA_CHANNEL = 'https://anaconda.org/qiime2'
from .method import Method from .plugin_manager import PluginManager from .provenance import Provenance from .visualizer import Visualizer from .result import Result, Artifact, Visualization from ..core.util import parse_type __all__ = ['Result', 'Artifact', 'Visualization', 'Method', 'Visualizer', 'PluginManager', 'Provenance', 'parse_type'] + + # Various URLs + CITATION = 'http://www.ncbi.nlm.nih.gov/pubmed/20383131' + HELP_URL = 'http://2.qiime.org' + CONDA_CHANNEL = 'https://anaconda.org/qiime2'
29fef644079a03fe0cfeb792dd47af7749382dba
unnaturalcode/http/__main__.py
unnaturalcode/http/__main__.py
from unnaturalcode.http import unnaturalhttp from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0')
try: from unnaturalcode.http import unnaturalhttp except ImportError: import sys, os # Oiugh. sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) from unnaturalcode.http import unnaturalhttp from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0')
Fix to allow invocation by `python unnaturalcode/http`
Fix to allow invocation by `python unnaturalcode/http`
Python
agpl-3.0
orezpraw/unnaturalcode,orezpraw/unnaturalcode,orezpraw/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,orezpraw/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,naturalness/unnaturalcode,orezpraw/estimate-charm,naturalness/unnaturalcode,orezpraw/unnaturalcode
+ + try: - from unnaturalcode.http import unnaturalhttp + from unnaturalcode.http import unnaturalhttp + except ImportError: + import sys, os + # Oiugh. + sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) + from unnaturalcode.http import unnaturalhttp + from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0')
Fix to allow invocation by `python unnaturalcode/http`
## Code Before: from unnaturalcode.http import unnaturalhttp from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0') ## Instruction: Fix to allow invocation by `python unnaturalcode/http` ## Code After: try: from unnaturalcode.http import unnaturalhttp except ImportError: import sys, os # Oiugh. sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) from unnaturalcode.http import unnaturalhttp from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0')
+ + try: - from unnaturalcode.http import unnaturalhttp + from unnaturalcode.http import unnaturalhttp ? ++++ + except ImportError: + import sys, os + # Oiugh. + sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) + from unnaturalcode.http import unnaturalhttp + from flask import Flask app = Flask(__name__) app.register_blueprint(unnaturalhttp) app.run(host='0.0.0.0')
edc773bfd5d25a42fa2759631500fc4861557e57
fireplace/cards/tgt/priest.py
fireplace/cards/tgt/priest.py
from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5)
from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) # Shadowfiend class AT_014: events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e")) # Wyrmrest Agent class AT_116: play = HOLDING_DRAGON & Buff(SELF, "AT_116e") ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5)
Implement more TGT Priest cards
Implement more TGT Priest cards
Python
agpl-3.0
oftc-ftw/fireplace,beheh/fireplace,smallnamespace/fireplace,Meerkov/fireplace,Ragowit/fireplace,smallnamespace/fireplace,oftc-ftw/fireplace,amw2104/fireplace,liujimj/fireplace,NightKev/fireplace,jleclanche/fireplace,Ragowit/fireplace,amw2104/fireplace,liujimj/fireplace,Meerkov/fireplace
from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) + + + # Shadowfiend + class AT_014: + events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e")) + + + # Wyrmrest Agent + class AT_116: + play = HOLDING_DRAGON & Buff(SELF, "AT_116e") ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5)
Implement more TGT Priest cards
## Code Before: from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5) ## Instruction: Implement more TGT Priest cards ## Code After: from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) # Shadowfiend class AT_014: events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e")) # Wyrmrest Agent class AT_116: play = HOLDING_DRAGON & Buff(SELF, "AT_116e") ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5)
from ..utils import * ## # Minions # Holy Champion class AT_011: events = Heal().on(Buff(SELF, "AT_011e")) # Spawn of Shadows class AT_012: inspire = Hit(ALL_HEROES, 4) + + + # Shadowfiend + class AT_014: + events = Draw(CONTROLLER).on(Buff(Draw.Args.CARD, "AT_014e")) + + + # Wyrmrest Agent + class AT_116: + play = HOLDING_DRAGON & Buff(SELF, "AT_116e") ## # Spells # Power Word: Glory class AT_013: play = Buff(TARGET, "AT_013e") class AT_013e: events = Attack(OWNER).on(Heal(FRIENDLY_HERO, 4)) # Convert class AT_015: play = Give(CONTROLLER, Copy(TARGET)) # Confuse class AT_016: play = Buff(ALL_MINIONS, "AT_016e") # Flash Heal class AT_055: play = Heal(TARGET, 5)
0b7f99bcb4e42c50263a7d8a42513876b02b445a
scikits/talkbox/tools/__init__.py
scikits/talkbox/tools/__init__.py
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter']
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
Put segment_axis in the main scikits.talkbox namespace.
Put segment_axis in the main scikits.talkbox namespace.
Python
mit
cournape/talkbox,cournape/talkbox
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] + from segmentaxis import segment_axis + __all__ += ['segment_axis'] +
Put segment_axis in the main scikits.talkbox namespace.
## Code Before: __all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] ## Instruction: Put segment_axis in the main scikits.talkbox namespace. ## Code After: __all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] from segmentaxis import segment_axis __all__ += ['segment_axis']
__all__ = [] import correlations from correlations import * __all__ += correlations.__all__ import cffilter from cffilter import cslfilter as slfilter __all__ += ['slfilter'] + + from segmentaxis import segment_axis + __all__ += ['segment_axis']
df8f5e0a6be5f3de31d61810b1624175b2d105ec
auth0/v2/device_credentials.py
auth0/v2/device_credentials.py
from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params)
from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params) def create(self, body): return self.client.post(data=body) def delete(self, id): return self.client.delete(id=id)
Implement create and delete methods for DeviceCredentials
Implement create and delete methods for DeviceCredentials
Python
mit
auth0/auth0-python,auth0/auth0-python
from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params) + def create(self, body): + return self.client.post(data=body) + + def delete(self, id): + return self.client.delete(id=id) +
Implement create and delete methods for DeviceCredentials
## Code Before: from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params) ## Instruction: Implement create and delete methods for DeviceCredentials ## Code After: from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params) def create(self, body): return self.client.post(data=body) def delete(self, id): return self.client.delete(id=id)
from .rest import RestClient class DeviceCredentials(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/device-credentials' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, user_id=None, client_id=None, type=None, fields=[], include_fields=True): params = { 'fields': ','.join(fields) or None, 'include_fields': str(include_fields).lower(), 'user_id': user_id, 'client_id': client_id, 'type': type, } return self.client.get(params=params) + + def create(self, body): + return self.client.post(data=body) + + def delete(self, id): + return self.client.delete(id=id)
278b17859e4ad7464098a715777fcb755acf258c
doTranscode.py
doTranscode.py
import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file f = tempfile.NamedTemporaryFile() inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ
import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file f = tempfile.NamedTemporaryFile(suffix=".wav") inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ
Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file
Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file
Python
isc
jeffayle/Transcode
import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file - f = tempfile.NamedTemporaryFile() + f = tempfile.NamedTemporaryFile(suffix=".wav") inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ
Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file
## Code Before: import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file f = tempfile.NamedTemporaryFile() inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ ## Instruction: Make sure that the temporary file has a `wav` extension because a certain encoder was designed for Windows and thinks that you would never possibly have a file without an extension so adds `.wav` if it's not there on the input file ## Code After: import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file f = tempfile.NamedTemporaryFile(suffix=".wav") inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ
import encoders import decoders import config import tempfile import os def transcode(inF, outF, options, type=None): "Transcodes a file" if type == None: type = os.path.splitext(outF)[1][1:].lower() #Get the file's metadata meta = decoders.getMetadata(inF) #Decode the file - f = tempfile.NamedTemporaryFile() + f = tempfile.NamedTemporaryFile(suffix=".wav") ? +++++++++++++ inF_real = decoders.decode(inF, f.name) if not inF_real: return False #Encode it succ = encoders.encode(inF_real, outF, type, options, meta) #Clean up f.close() return succ
0a13a9a8a779102dbcb2beead7d8aa9143f4c79b
tests/pytests/unit/client/ssh/test_shell.py
tests/pytests/unit/client/ssh/test_shell.py
import os import subprocess import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" yield {"pub_key": str(pub_key), "priv_key": str(priv_key)} @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) class TestSSHShell: def test_ssh_shell_key_gen(self, keys): """ Test ssh key_gen """ shell.gen_key(keys["priv_key"]) for fp in keys.keys(): assert os.path.exists(keys[fp]) # verify there is not a passphrase set on key ret = subprocess.check_output( ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30, ) assert ret.decode().startswith("ssh-rsa")
import subprocess import types import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key) @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) def test_ssh_shell_key_gen(keys): """ Test ssh key_gen """ shell.gen_key(str(keys.priv_key)) assert keys.priv_key.exists() assert keys.pub_key.exists() # verify there is not a passphrase set on key ret = subprocess.check_output( ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30, ) assert ret.decode().startswith("ssh-rsa")
Use commit suggestion to use types
Use commit suggestion to use types Co-authored-by: Pedro Algarvio <4410d99cefe57ec2c2cdbd3f1d5cf862bb4fb6f8@algarvio.me>
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
- import os import subprocess + import types import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" - yield {"pub_key": str(pub_key), "priv_key": str(priv_key)} + return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key) @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) - class TestSSHShell: - def test_ssh_shell_key_gen(self, keys): + def test_ssh_shell_key_gen(keys): - """ + """ - Test ssh key_gen + Test ssh key_gen - """ + """ - shell.gen_key(keys["priv_key"]) + shell.gen_key(str(keys.priv_key)) - for fp in keys.keys(): - assert os.path.exists(keys[fp]) + assert keys.priv_key.exists() + assert keys.pub_key.exists() + # verify there is not a passphrase set on key + ret = subprocess.check_output( + ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30, + ) + assert ret.decode().startswith("ssh-rsa") - # verify there is not a passphrase set on key - ret = subprocess.check_output( - ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30, - ) - assert ret.decode().startswith("ssh-rsa") -
Use commit suggestion to use types
## Code Before: import os import subprocess import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" yield {"pub_key": str(pub_key), "priv_key": str(priv_key)} @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) class TestSSHShell: def test_ssh_shell_key_gen(self, keys): """ Test ssh key_gen """ shell.gen_key(keys["priv_key"]) for fp in keys.keys(): assert os.path.exists(keys[fp]) # verify there is not a passphrase set on key ret = subprocess.check_output( ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30, ) assert ret.decode().startswith("ssh-rsa") ## Instruction: Use commit suggestion to use types ## Code After: import subprocess import types import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key) @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) def test_ssh_shell_key_gen(keys): """ Test ssh key_gen """ shell.gen_key(str(keys.priv_key)) assert keys.priv_key.exists() assert keys.pub_key.exists() # verify there is not a passphrase set on key ret = subprocess.check_output( ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30, ) assert ret.decode().startswith("ssh-rsa")
- import os import subprocess + import types import pytest import salt.client.ssh.shell as shell @pytest.fixture def keys(tmp_path): pub_key = tmp_path / "ssh" / "testkey.pub" priv_key = tmp_path / "ssh" / "testkey" - yield {"pub_key": str(pub_key), "priv_key": str(priv_key)} + return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key) @pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh") @pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True) - class TestSSHShell: - def test_ssh_shell_key_gen(self, keys): ? ---- ------ + def test_ssh_shell_key_gen(keys): - """ ? ---- + """ - Test ssh key_gen ? ---- + Test ssh key_gen - """ ? ---- + """ - shell.gen_key(keys["priv_key"]) ? ---- ^^ -- + shell.gen_key(str(keys.priv_key)) ? ++++ ^ + + assert keys.priv_key.exists() + assert keys.pub_key.exists() - for fp in keys.keys(): - assert os.path.exists(keys[fp]) - - # verify there is not a passphrase set on key ? ---- + # verify there is not a passphrase set on key - ret = subprocess.check_output( ? ---- + ret = subprocess.check_output( - ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30, ? ---- ^^ ^^ + ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30, ? ++++ ^ ^ - ) + ) - assert ret.decode().startswith("ssh-rsa") ? ---- + assert ret.decode().startswith("ssh-rsa")
c668aaa0f22f5a61094c2028291b65c781733a54
mojapi/api.py
mojapi/api.py
import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json()
import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json() def get_blocked_server_hashes(): response = requests.get('https://sessionserver.mojang.com/blockedservers') response.raise_for_status() sha1_hashes = response.content.split(b'\n') return sha1_hashes
Add get blocked server hashes call
Add get blocked server hashes call
Python
mit
zugmc/mojapi
import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json() + + def get_blocked_server_hashes(): + response = requests.get('https://sessionserver.mojang.com/blockedservers') + response.raise_for_status() + sha1_hashes = response.content.split(b'\n') + return sha1_hashes +
Add get blocked server hashes call
## Code Before: import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json() ## Instruction: Add get blocked server hashes call ## Code After: import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json() def get_blocked_server_hashes(): response = requests.get('https://sessionserver.mojang.com/blockedservers') response.raise_for_status() sha1_hashes = response.content.split(b'\n') return sha1_hashes
import json import requests import time def get_statuses(): return requests.get('https://status.mojang.com/check/').json() def get_uuid(username, unix_timestamp=None): if unix_timestamp is None: unix_timestamp = int(time.time()) return requests.get( 'https://api.mojang.com/users/profiles/minecraft/{}?at={}'.format(username, unix_timestamp) ).json() def get_usernames(uuid): return requests.get('https://api.mojang.com/user/profiles/{}/names'.format(uuid)).json() def get_profiles(*usernames): return requests.post( url='https://api.mojang.com/profiles/minecraft', headers={ b'Content-Type': b'application/json' }, data=json.dumps(list(usernames)) ).json() + + + def get_blocked_server_hashes(): + response = requests.get('https://sessionserver.mojang.com/blockedservers') + response.raise_for_status() + sha1_hashes = response.content.split(b'\n') + return sha1_hashes
5dec1db567ef7c2b6ea1cca3ddd02612cb9f7d8a
Lib/encodings/bz2_codec.py
Lib/encodings/bz2_codec.py
import codecs import bz2 def encode(input, errors='strict'): assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) def decode(input, errors='strict'): assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) ### encodings module API def getregentry(): return (encode, decode, codecs.StreamReader, codecs.StreamWriter)
import codecs import bz2 # this codec needs the optional bz2 module ! ### Codec APIs def bz2_encode(input,errors='strict'): """ Encodes the object input and returns a tuple (output object, length consumed). errors defines the error handling to apply. It defaults to 'strict' handling which is the only currently supported error handling for this codec. """ assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) def bz2_decode(input,errors='strict'): """ Decodes the object input and returns a tuple (output object, length consumed). input must be an object which provides the bf_getreadbuf buffer slot. Python strings, buffer objects and memory mapped files are examples of objects providing this slot. errors defines the error handling to apply. It defaults to 'strict' handling which is the only currently supported error handling for this codec. """ assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) class Codec(codecs.Codec): def encode(self, input, errors='strict'): return bz2_encode(input, errors) def decode(self, input, errors='strict'): return bz2_decode(input, errors) class StreamWriter(Codec,codecs.StreamWriter): pass class StreamReader(Codec,codecs.StreamReader): pass ### encodings module API def getregentry(): return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
Revert previous change. MAL preferred the old version.
Revert previous change. MAL preferred the old version.
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
import codecs - import bz2 + import bz2 # this codec needs the optional bz2 module ! + ### Codec APIs + - def encode(input, errors='strict'): + def bz2_encode(input,errors='strict'): + + """ Encodes the object input and returns a tuple (output + object, length consumed). + + errors defines the error handling to apply. It defaults to + 'strict' handling which is the only currently supported + error handling for this codec. + + """ assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) - def decode(input, errors='strict'): + def bz2_decode(input,errors='strict'): + + """ Decodes the object input and returns a tuple (output + object, length consumed). + + input must be an object which provides the bf_getreadbuf + buffer slot. Python strings, buffer objects and memory + mapped files are examples of objects providing this slot. + + errors defines the error handling to apply. It defaults to + 'strict' handling which is the only currently supported + error handling for this codec. + + """ assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) + + class Codec(codecs.Codec): + + def encode(self, input, errors='strict'): + return bz2_encode(input, errors) + def decode(self, input, errors='strict'): + return bz2_decode(input, errors) + + class StreamWriter(Codec,codecs.StreamWriter): + pass + + class StreamReader(Codec,codecs.StreamReader): + pass ### encodings module API def getregentry(): - return (encode, decode, codecs.StreamReader, codecs.StreamWriter) + return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
Revert previous change. MAL preferred the old version.
## Code Before: import codecs import bz2 def encode(input, errors='strict'): assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) def decode(input, errors='strict'): assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) ### encodings module API def getregentry(): return (encode, decode, codecs.StreamReader, codecs.StreamWriter) ## Instruction: Revert previous change. MAL preferred the old version. ## Code After: import codecs import bz2 # this codec needs the optional bz2 module ! ### Codec APIs def bz2_encode(input,errors='strict'): """ Encodes the object input and returns a tuple (output object, length consumed). errors defines the error handling to apply. It defaults to 'strict' handling which is the only currently supported error handling for this codec. """ assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) def bz2_decode(input,errors='strict'): """ Decodes the object input and returns a tuple (output object, length consumed). input must be an object which provides the bf_getreadbuf buffer slot. Python strings, buffer objects and memory mapped files are examples of objects providing this slot. errors defines the error handling to apply. It defaults to 'strict' handling which is the only currently supported error handling for this codec. """ assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) class Codec(codecs.Codec): def encode(self, input, errors='strict'): return bz2_encode(input, errors) def decode(self, input, errors='strict'): return bz2_decode(input, errors) class StreamWriter(Codec,codecs.StreamWriter): pass class StreamReader(Codec,codecs.StreamReader): pass ### encodings module API def getregentry(): return (bz2_encode,bz2_decode,StreamReader,StreamWriter)
import codecs - import bz2 + import bz2 # this codec needs the optional bz2 module ! + ### Codec APIs + - def encode(input, errors='strict'): ? - + def bz2_encode(input,errors='strict'): ? ++++ + + """ Encodes the object input and returns a tuple (output + object, length consumed). + + errors defines the error handling to apply. It defaults to + 'strict' handling which is the only currently supported + error handling for this codec. + + """ assert errors == 'strict' output = bz2.compress(input) return (output, len(input)) - def decode(input, errors='strict'): ? - + def bz2_decode(input,errors='strict'): ? ++++ + + """ Decodes the object input and returns a tuple (output + object, length consumed). + + input must be an object which provides the bf_getreadbuf + buffer slot. Python strings, buffer objects and memory + mapped files are examples of objects providing this slot. + + errors defines the error handling to apply. It defaults to + 'strict' handling which is the only currently supported + error handling for this codec. + + """ assert errors == 'strict' output = bz2.decompress(input) return (output, len(input)) + + class Codec(codecs.Codec): + + def encode(self, input, errors='strict'): + return bz2_encode(input, errors) + def decode(self, input, errors='strict'): + return bz2_decode(input, errors) + + class StreamWriter(Codec,codecs.StreamWriter): + pass + + class StreamReader(Codec,codecs.StreamReader): + pass ### encodings module API def getregentry(): - return (encode, decode, codecs.StreamReader, codecs.StreamWriter) ? ^ -------- -------- + return (bz2_encode,bz2_decode,StreamReader,StreamWriter) ? ++++ ^^^^
f22476a36f2096628dc336f9adf0caa9a827dc11
jfr_playoff/db.py
jfr_playoff/db.py
import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e)) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e))
import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: raise IOError(e.errno, str(e), db_name) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e))
Fix for rethrowing mysql.connector.Error as IOError
Fix for rethrowing mysql.connector.Error as IOError
Python
bsd-2-clause
emkael/jfrteamy-playoff,emkael/jfrteamy-playoff
import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: + raise IOError(e.errno, str(e), db_name) - raise IOError( - message=str(e), filename=db_name, - errno=e.errno, strerror=str(e)) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e))
Fix for rethrowing mysql.connector.Error as IOError
## Code Before: import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e)) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e)) ## Instruction: Fix for rethrowing mysql.connector.Error as IOError ## Code After: import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: raise IOError(e.errno, str(e), db_name) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e))
import sys class PlayoffDB(object): db_cursor = None DATABASE_NOT_CONFIGURED_WARNING = 'WARNING: database not configured' def __init__(self, settings): reload(sys) sys.setdefaultencoding("latin1") import mysql.connector self.database = mysql.connector.connect( user=settings['user'], password=settings['pass'], host=settings['host'], port=settings['port']) self.db_cursor = self.database.cursor(buffered=True) def get_cursor(self): return self.db_cursor def __execute_query(self, db_name, sql, params): self.db_cursor.execute(sql.replace('#db#', db_name), params) def fetch(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) row = self.db_cursor.fetchone() return row except mysql.connector.Error as e: + raise IOError(e.errno, str(e), db_name) - raise IOError( - message=str(e), filename=db_name, - errno=e.errno, strerror=str(e)) def fetch_all(self, db_name, sql, params): import mysql.connector try: self.__execute_query(db_name, sql, params) results = self.db_cursor.fetchall() return results except mysql.connector.Error as e: raise IOError( message=str(e), filename=db_name, errno=e.errno, strerror=str(e))
7917716ebd11770c5d4d0634b39e32e4f577ab71
tests/test_urls.py
tests/test_urls.py
from unittest import TestCase class TestURLs(TestCase): pass
from unittest import TestCase from django.contrib.auth import views from django.core.urlresolvers import resolve, reverse class URLsMixin(object): """ A TestCase Mixin with a check_url helper method for testing urls. Pirated with slight modifications from incuna_test_utils https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py """ def check_url(self, view_method, expected_url, url_name, url_args=None, url_kwargs=None): """ Assert a view's url is correctly configured Check the url_name reverses to give a correctly formated expected_url. Check the expected_url resolves to the correct view. """ reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs) self.assertEqual(reversed_url, expected_url) # Look for a method rather than a class here # (just because of what we're testing) resolved_view_method = resolve(expected_url).func self.assertEqual(resolved_view_method, view_method) class TestURLs(URLsMixin, TestCase): def test_login(self): self.check_url( views.login, '/login/', 'login', ) def test_logout(self): self.check_url( views.logout, '/logout/', 'logout', ) def test_password_change(self): self.check_url( views.password_change, '/password/change/', 'password_change', ) def test_password_change_done(self): self.check_url( views.password_change_done, '/password/change/done/', 'password_change_done', ) def test_password_reset(self): self.check_url( views.password_reset, '/password/reset/', 'password_reset', ) def test_password_reset_done(self): self.check_url( views.password_reset_done, '/password/reset/done/', 'password_reset_done', ) def test_password_reset_complete(self): self.check_url( views.password_reset_complete, '/password/reset/complete/', 'password_reset_complete', )
Add lots of URL tests.
Add lots of URL tests. * The URLsMixin from incuna_test_utils/testcases/urls.py isn't quite doing what we want here, so rip it off and make a small modification (resolve(...).func.cls -> resolve(...).func). * Add lots of tests for the django.contrib.auth views that we're using (the others are more complex).
Python
bsd-2-clause
incuna/incuna-auth,ghickman/incuna-auth,ghickman/incuna-auth,incuna/incuna-auth
from unittest import TestCase + from django.contrib.auth import views + from django.core.urlresolvers import resolve, reverse - class TestURLs(TestCase): - pass + class URLsMixin(object): + """ + A TestCase Mixin with a check_url helper method for testing urls. + Pirated with slight modifications from incuna_test_utils + https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py + """ + + def check_url(self, view_method, expected_url, url_name, + url_args=None, url_kwargs=None): + """ + Assert a view's url is correctly configured + + Check the url_name reverses to give a correctly formated expected_url. + Check the expected_url resolves to the correct view. + """ + + reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs) + self.assertEqual(reversed_url, expected_url) + + # Look for a method rather than a class here + # (just because of what we're testing) + resolved_view_method = resolve(expected_url).func + self.assertEqual(resolved_view_method, view_method) + + + class TestURLs(URLsMixin, TestCase): + + def test_login(self): + self.check_url( + views.login, + '/login/', + 'login', + ) + + def test_logout(self): + self.check_url( + views.logout, + '/logout/', + 'logout', + ) + + def test_password_change(self): + self.check_url( + views.password_change, + '/password/change/', + 'password_change', + ) + + def test_password_change_done(self): + self.check_url( + views.password_change_done, + '/password/change/done/', + 'password_change_done', + ) + + def test_password_reset(self): + self.check_url( + views.password_reset, + '/password/reset/', + 'password_reset', + ) + + def test_password_reset_done(self): + self.check_url( + views.password_reset_done, + '/password/reset/done/', + 'password_reset_done', + ) + + def test_password_reset_complete(self): + self.check_url( + views.password_reset_complete, + '/password/reset/complete/', + 'password_reset_complete', + ) +
Add lots of URL tests.
## Code Before: from unittest import TestCase class TestURLs(TestCase): pass ## Instruction: Add lots of URL tests. ## Code After: from unittest import TestCase from django.contrib.auth import views from django.core.urlresolvers import resolve, reverse class URLsMixin(object): """ A TestCase Mixin with a check_url helper method for testing urls. Pirated with slight modifications from incuna_test_utils https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py """ def check_url(self, view_method, expected_url, url_name, url_args=None, url_kwargs=None): """ Assert a view's url is correctly configured Check the url_name reverses to give a correctly formated expected_url. Check the expected_url resolves to the correct view. """ reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs) self.assertEqual(reversed_url, expected_url) # Look for a method rather than a class here # (just because of what we're testing) resolved_view_method = resolve(expected_url).func self.assertEqual(resolved_view_method, view_method) class TestURLs(URLsMixin, TestCase): def test_login(self): self.check_url( views.login, '/login/', 'login', ) def test_logout(self): self.check_url( views.logout, '/logout/', 'logout', ) def test_password_change(self): self.check_url( views.password_change, '/password/change/', 'password_change', ) def test_password_change_done(self): self.check_url( views.password_change_done, '/password/change/done/', 'password_change_done', ) def test_password_reset(self): self.check_url( views.password_reset, '/password/reset/', 'password_reset', ) def test_password_reset_done(self): self.check_url( views.password_reset_done, '/password/reset/done/', 'password_reset_done', ) def test_password_reset_complete(self): self.check_url( views.password_reset_complete, '/password/reset/complete/', 'password_reset_complete', )
from unittest import TestCase + from django.contrib.auth import views + from django.core.urlresolvers import resolve, reverse + + class URLsMixin(object): + """ + A TestCase Mixin with a check_url helper method for testing urls. + Pirated with slight modifications from incuna_test_utils + https://github.com/incuna/incuna-test-utils/blob/master/incuna_test_utils/testcases/urls.py + """ + + def check_url(self, view_method, expected_url, url_name, + url_args=None, url_kwargs=None): + """ + Assert a view's url is correctly configured + + Check the url_name reverses to give a correctly formated expected_url. + Check the expected_url resolves to the correct view. + """ + + reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs) + self.assertEqual(reversed_url, expected_url) + + # Look for a method rather than a class here + # (just because of what we're testing) + resolved_view_method = resolve(expected_url).func + self.assertEqual(resolved_view_method, view_method) + + - class TestURLs(TestCase): + class TestURLs(URLsMixin, TestCase): ? +++++++++++ - pass + + def test_login(self): + self.check_url( + views.login, + '/login/', + 'login', + ) + + def test_logout(self): + self.check_url( + views.logout, + '/logout/', + 'logout', + ) + + def test_password_change(self): + self.check_url( + views.password_change, + '/password/change/', + 'password_change', + ) + + def test_password_change_done(self): + self.check_url( + views.password_change_done, + '/password/change/done/', + 'password_change_done', + ) + + def test_password_reset(self): + self.check_url( + views.password_reset, + '/password/reset/', + 'password_reset', + ) + + def test_password_reset_done(self): + self.check_url( + views.password_reset_done, + '/password/reset/done/', + 'password_reset_done', + ) + + def test_password_reset_complete(self): + self.check_url( + views.password_reset_complete, + '/password/reset/complete/', + 'password_reset_complete', + )
6fecc53b63023e6d25722aa66038285be3b4d46b
arcutils/response.py
arcutils/response.py
from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() location = ( request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) or request.META.get('HTTP_REFERER') ) if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default return location
from urllib.parse import urlparse, urlunparse from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) if location: from_referrer = False else: location = request.META.get('HTTP_REFERER') from_referrer = bool(location) if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default elif from_referrer: info = urlparse(location) if info.netloc == host: # Clear scheme and host (AKA netloc) to get just the path of # the referrer. Also, ensure the path is set for consistency. new_info = ('', '', info.path or '/') + info[3:] location = urlunparse(new_info) return location
Return just path when getting redirect location from REFERER
Return just path when getting redirect location from REFERER In response.get_redirect_location(). There's no need to include the scheme and host in a redirect back to the same site. Removing them makes redirect URLs more concise.
Python
mit
PSU-OIT-ARC/django-arcutils,wylee/django-arcutils,PSU-OIT-ARC/django-arcutils,wylee/django-arcutils
+ from urllib.parse import urlparse, urlunparse + from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() + location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) + - location = ( + if location: - request.GET.get(redirect_field_name) or - request.POST.get(redirect_field_name) or + from_referrer = False + else: - request.META.get('HTTP_REFERER') + location = request.META.get('HTTP_REFERER') - ) + from_referrer = bool(location) + if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default + elif from_referrer: + info = urlparse(location) + if info.netloc == host: + # Clear scheme and host (AKA netloc) to get just the path of + # the referrer. Also, ensure the path is set for consistency. + new_info = ('', '', info.path or '/') + info[3:] + location = urlunparse(new_info) + return location
Return just path when getting redirect location from REFERER
## Code Before: from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() location = ( request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) or request.META.get('HTTP_REFERER') ) if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default return location ## Instruction: Return just path when getting redirect location from REFERER ## Code After: from urllib.parse import urlparse, urlunparse from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) if location: from_referrer = False else: location = request.META.get('HTTP_REFERER') from_referrer = bool(location) if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default elif from_referrer: info = urlparse(location) if info.netloc == host: # Clear scheme and host (AKA netloc) to get just the path of # the referrer. Also, ensure the path is set for consistency. new_info = ('', '', info.path or '/') + info[3:] location = urlunparse(new_info) return location
+ from urllib.parse import urlparse, urlunparse + from django.contrib.auth import REDIRECT_FIELD_NAME from django.utils.http import is_safe_url def get_redirect_location(request, redirect_field_name=REDIRECT_FIELD_NAME, default='/'): """Attempt to choose an optimal redirect location. If a location is specified via a request parameter, that location will be used. If a location is specified via POST or PUT data, that location will be used. In either of the above two cases, the name of the parameter and data field is specified by ``redirect_field_name``, which defaults to "next". Otherwise, the preferred option is to redirect back to the referring page. If there's no referrer, the default is used. In any case, the redirect location must be safe (same host, safe scheme). Otherwise, the ``default`` location will be used. If the default location isn't safe, "/" will be used as a last resort. """ host = request.get_host() + location = request.GET.get(redirect_field_name) or request.POST.get(redirect_field_name) + - location = ( ? ^^^^ + if location: ? +++ ^ - request.GET.get(redirect_field_name) or - request.POST.get(redirect_field_name) or + from_referrer = False + else: - request.META.get('HTTP_REFERER') + location = request.META.get('HTTP_REFERER') ? +++++++++++ - ) + from_referrer = bool(location) + if not is_safe_url(location, host): default = default or '/' if not is_safe_url(default, host): default = '/' location = default + elif from_referrer: + info = urlparse(location) + if info.netloc == host: + # Clear scheme and host (AKA netloc) to get just the path of + # the referrer. Also, ensure the path is set for consistency. + new_info = ('', '', info.path or '/') + info[3:] + location = urlunparse(new_info) + return location
456b72757cda81c8dd6634ae41b8a1008ff59087
config-example.py
config-example.py
# put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ]
# put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) # Use this picture as a watermark, for sponsorships etc. Should be a PNG image # with transparency. It is overlaid directly with the camera image, so it # should have the same dimensions as `camera_dimensions` above. Leave as an # empty string to have no watermark. watermark_path = "" # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ]
Add watermark path to example config
Add watermark path to example config
Python
mit
mgunyho/kiltiskahvi
# put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) + + # Use this picture as a watermark, for sponsorships etc. Should be a PNG image + # with transparency. It is overlaid directly with the camera image, so it + # should have the same dimensions as `camera_dimensions` above. Leave as an + # empty string to have no watermark. + watermark_path = "" + + # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ]
Add watermark path to example config
## Code Before: # put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ] ## Instruction: Add watermark path to example config ## Code After: # put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) # Use this picture as a watermark, for sponsorships etc. Should be a PNG image # with transparency. It is overlaid directly with the camera image, so it # should have the same dimensions as `camera_dimensions` above. Leave as an # empty string to have no watermark. watermark_path = "" # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ]
# put your bot token here as a string bot_token = "" # the tg username of the bot's admin. admin_username = "" # The size of the pictures the webcamera takes. As of 2022-03-06, the guild # room has a Creative Live! Cam Sync HD USB webcamera, which at least claims to # be 720p camera_dimensions = (1280, 720) + + # Use this picture as a watermark, for sponsorships etc. Should be a PNG image + # with transparency. It is overlaid directly with the camera image, so it + # should have the same dimensions as `camera_dimensions` above. Leave as an + # empty string to have no watermark. + watermark_path = "" + + # if a message contains any of these words, the bot responds trigger_words = [ "kahvi", "\u2615", # coffee emoji "tsufe", "kahavi", #"sima", # wappu mode ]
59ce3ca9c1572dcf71aa5de5cdb354def594a36c
downloads/urls.py
downloads/urls.py
from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', url(r'solutions/(?P<path>.*)$', partial(download_protected_file, path_prefix='solutions/', model_class=UserSolution), name='download_solution'), )
from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', url(r'solutions/(?P<path>.*)$', download_protected_file, dict(path_prefix='solutions/', model_class=UserSolution), name='download_solution'), )
Remove unnecessary usage of functools.partial
downloads: Remove unnecessary usage of functools.partial
Python
mit
matus-stehlik/roots,rtrembecky/roots,matus-stehlik/roots,tbabej/roots,rtrembecky/roots,tbabej/roots,tbabej/roots,rtrembecky/roots,matus-stehlik/roots
from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', - url(r'solutions/(?P<path>.*)$', partial(download_protected_file, + url(r'solutions/(?P<path>.*)$', download_protected_file, + dict(path_prefix='solutions/', model_class=UserSolution), - path_prefix='solutions/', - model_class=UserSolution), name='download_solution'), )
Remove unnecessary usage of functools.partial
## Code Before: from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', url(r'solutions/(?P<path>.*)$', partial(download_protected_file, path_prefix='solutions/', model_class=UserSolution), name='download_solution'), ) ## Instruction: Remove unnecessary usage of functools.partial ## Code After: from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', url(r'solutions/(?P<path>.*)$', download_protected_file, dict(path_prefix='solutions/', model_class=UserSolution), name='download_solution'), )
from django.conf.urls import patterns, url from functools import partial from problems.models import UserSolution from .views import download_protected_file urlpatterns = patterns('', - url(r'solutions/(?P<path>.*)$', partial(download_protected_file, ? -------- + url(r'solutions/(?P<path>.*)$', download_protected_file, + dict(path_prefix='solutions/', model_class=UserSolution), - path_prefix='solutions/', - model_class=UserSolution), name='download_solution'), )
59afb96f2211983ee2a2786c60791074b13c3e7f
ni/__main__.py
ni/__main__.py
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
Tweak comment about 202 response
Tweak comment about 202 response
Python
apache-2.0
python/the-knights-who-say-ni,python/the-knights-who-say-ni
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented - # With a background queue, one could add the update as a work item + # With a work queue, one could make the updating of the - # and return an HTTP 202 response. + # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
Tweak comment about 202 response
## Code Before: """Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a background queue, one could add the update as a work item # and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port()) ## Instruction: Tweak comment about 202 response ## Code After: """Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented # With a work queue, one could make the updating of the # contribution a work item and return an HTTP 202 response. return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
"""Implement a server to check if a contribution is covered by a CLA(s).""" from aiohttp import web from . import abc from . import ContribHost from . import ServerHost from . import CLAHost class Handler: """Handle requests from the contribution host.""" def __init__(self, server: ServerHost, cla_records: CLAHost): self.server = server self.cla_records = cla_records async def respond(request: web.Request) -> web.StreamResponse: # XXX untested """Handle a webhook trigger from the contribution host.""" try: contribution = ContribHost.process(request) usernames = await contribution.usernames() # XXX not implemented cla_status = await self.cla_records.check(usernames) # XXX not implemented - # With a background queue, one could add the update as a work item + # With a work queue, one could make the updating of the - # and return an HTTP 202 response. + # contribution a work item and return an HTTP 202 response. ? +++++++++++++++++++++++++ return (await contribution.update(cla_status)) # XXX not implemented except abc.ResponseExit as exc: return exc.response except Exception as exc: self.server.log(exc) return web.Response( status=http.HTTPStatus.INTERNAL_SERVER_ERROR.value) if __name__ == '__main__': server = ServerHost() cla_records = CLAHost() handler = Handler(server, cla_records) app = web.Application() app.router.add_route(*ContribHost.route, handler.respond) web.run_app(app, port=server.port())
2814d7b8060d1f468bb6fb34d1460cdad1811031
tools/android/emulator/reporting.py
tools/android/emulator/reporting.py
"""An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, unused_runtime_ms, unused_success): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter()
"""An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, unused_runtime_ms, unused_success, unused_total_runtime): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter()
Update the reporter interface to even track the total runtime
Update the reporter interface to even track the total runtime PiperOrigin-RevId: 160982468
Python
apache-2.0
android/android-test,android/android-test,android/android-test,android/android-test,android/android-test
"""An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, - unused_runtime_ms, unused_success): + unused_runtime_ms, unused_success, + unused_total_runtime): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter()
Update the reporter interface to even track the total runtime
## Code Before: """An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, unused_runtime_ms, unused_success): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter() ## Instruction: Update the reporter interface to even track the total runtime ## Code After: """An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, unused_runtime_ms, unused_success, unused_total_runtime): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter()
"""An interface to report the status of emulator launches.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import logging import os import uuid class NoOpReporter(object): """Captures all device and failure data and throws it away.""" def __init__(self): pass def ReportDeviceProperties(self, unused_emu_type, unused_props): pass def ReportFailure(self, unused_component, unused_details): pass def ReportToolsUsage(self, unused_namespace, unused_tool_name, - unused_runtime_ms, unused_success): ? ^^ + unused_runtime_ms, unused_success, ? ^ + unused_total_runtime): pass def Emit(self): pass def MakeReporter(): """Creates a reporter instance.""" return NoOpReporter()
a85beb35d7296b0a8bd5a385b44fa13fb9f178ed
imgur-clean.py
imgur-clean.py
import hashlib import re import os import sys IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)') def get_hash(fn): with open(fn, 'rb') as fh: hashsum = hashlib.md5(fh.read()).digest() return hashsum if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) sums = {} for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue new_fn = f'{match.group(1)}.{match.group(2)}' if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) hashsum = get_hash(new_fn) files = sums.get(hashsum, []) files.append(new_fn) sums[hashsum] = files for hashsum, files in sums.items(): if len(files) > 1: files_quoted = [f"'{x}'" for x in files] print(f"Found duplicates: {', '.join(files_quoted)}") files.sort() for fn in files[1:]: os.remove(fn)
import re import os import sys IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)') if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) ids = {} for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue new_fn = f'{match[1]}.{match[3]}' if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) id = match[2] files = ids.get(id, []) files.append(new_fn) ids[id] = files for _, files in ids.items(): if len(files) > 1: files_quoted = ', '.join(f"'{fn}'" for fn in files) print(f"Found duplicates: {files_quoted}") files.sort() for fn in files[1:]: print(f"Removing {fn}") os.remove(fn)
Remove imgur duplicates based on ID.
Remove imgur duplicates based on ID.
Python
mit
ammongit/scripts,ammongit/scripts,ammongit/scripts,ammongit/scripts
- import hashlib import re import os import sys - IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)') + IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)') - - def get_hash(fn): - with open(fn, 'rb') as fh: - hashsum = hashlib.md5(fh.read()).digest() - return hashsum if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) - sums = {} + ids = {} for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue - new_fn = f'{match.group(1)}.{match.group(2)}' + new_fn = f'{match[1]}.{match[3]}' if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) - hashsum = get_hash(new_fn) + id = match[2] - files = sums.get(hashsum, []) + files = ids.get(id, []) files.append(new_fn) - sums[hashsum] = files + ids[id] = files - for hashsum, files in sums.items(): + for _, files in ids.items(): if len(files) > 1: - files_quoted = [f"'{x}'" for x in files] + files_quoted = ', '.join(f"'{fn}'" for fn in files) - print(f"Found duplicates: {', '.join(files_quoted)}") + print(f"Found duplicates: {files_quoted}") files.sort() for fn in files[1:]: + print(f"Removing {fn}") os.remove(fn)
Remove imgur duplicates based on ID.
## Code Before: import hashlib import re import os import sys IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)') def get_hash(fn): with open(fn, 'rb') as fh: hashsum = hashlib.md5(fh.read()).digest() return hashsum if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) sums = {} for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue new_fn = f'{match.group(1)}.{match.group(2)}' if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) hashsum = get_hash(new_fn) files = sums.get(hashsum, []) files.append(new_fn) sums[hashsum] = files for hashsum, files in sums.items(): if len(files) > 1: files_quoted = [f"'{x}'" for x in files] print(f"Found duplicates: {', '.join(files_quoted)}") files.sort() for fn in files[1:]: os.remove(fn) ## Instruction: Remove imgur duplicates based on ID. ## Code After: import re import os import sys IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)') if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) ids = {} for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue new_fn = f'{match[1]}.{match[3]}' if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) id = match[2] files = ids.get(id, []) files.append(new_fn) ids[id] = files for _, files in ids.items(): if len(files) > 1: files_quoted = ', '.join(f"'{fn}'" for fn in files) print(f"Found duplicates: {files_quoted}") files.sort() for fn in files[1:]: print(f"Removing {fn}") os.remove(fn)
- import hashlib import re import os import sys - IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)(?:-\w+)?\.([A-Za-z0-9]+)') ? --- - + IMGUR_FILENAME_REGEX = re.compile(r'([0-9]+)-(\w+)\.([A-Za-z0-9]+)') ? + - - def get_hash(fn): - with open(fn, 'rb') as fh: - hashsum = hashlib.md5(fh.read()).digest() - return hashsum if __name__ == '__main__': if len(sys.argv) >= 2: os.chdir(sys.argv[1]) - sums = {} ? ^^^ + ids = {} ? ^^ for fn in os.listdir('.'): match = IMGUR_FILENAME_REGEX.match(fn) if match is None: continue - new_fn = f'{match.group(1)}.{match.group(2)}' ? ^^^^^^^ ^ ^^^^^^^^^ + new_fn = f'{match[1]}.{match[3]}' ? ^ ^ ^^^ if fn == new_fn: continue print(f"Renaming '{fn}' to '{new_fn}'") os.rename(fn, new_fn) - hashsum = get_hash(new_fn) + id = match[2] - files = sums.get(hashsum, []) ? ^^^ ^^^^^^^ + files = ids.get(id, []) ? ^^ ^^ files.append(new_fn) - sums[hashsum] = files + ids[id] = files - for hashsum, files in sums.items(): ? ^^^^^^^ ^^^ + for _, files in ids.items(): ? ^ ^^ if len(files) > 1: - files_quoted = [f"'{x}'" for x in files] ? ^ ^ ^ ^ + files_quoted = ', '.join(f"'{fn}'" for fn in files) ? ^^^^^^^^^^ ^^ ^^ ^ - print(f"Found duplicates: {', '.join(files_quoted)}") ? ---------- - + print(f"Found duplicates: {files_quoted}") files.sort() for fn in files[1:]: + print(f"Removing {fn}") os.remove(fn)
efe1417ad049e4bb78bf1f111db6b2ea9c603461
rapt/util.py
rapt/util.py
import sys import yaml import click def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip()
import sys import yaml import click def load_yaml(fh_or_string): return yaml.safe_load(fh_or_string) def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip()
Add a load yaml helper
Add a load yaml helper
Python
bsd-3-clause
yougov/rapt,yougov/rapt
import sys import yaml import click + + + def load_yaml(fh_or_string): + return yaml.safe_load(fh_or_string) def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip()
Add a load yaml helper
## Code Before: import sys import yaml import click def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip() ## Instruction: Add a load yaml helper ## Code After: import sys import yaml import click def load_yaml(fh_or_string): return yaml.safe_load(fh_or_string) def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip()
import sys import yaml import click + + + def load_yaml(fh_or_string): + return yaml.safe_load(fh_or_string) def dump_yaml(obj): return yaml.dump(obj, default_flow_style=False) def edit_yaml(content='', footer=''): MARKER = '# Everything below is ignored\n\n' message = click.edit(content + '\n\n' + MARKER + footer, extension='.yaml') if message is not None: yaml_content = message.split(MARKER, 1)[0].rstrip('\n') return yaml.safe_load(yaml_content) def stdin(): for line in sys.stdin: yield line.strip()
6a15b33d69d8d66643bb8886f9916fa28ecaedea
molo/yourwords/templatetags/competition_tag.py
molo/yourwords/templatetags/competition_tag.py
from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: competitions = [] context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return []
from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: competitions = YourWordsCompetition.objects.none() context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return []
Return None if there is no competition
Return None if there is no competition
Python
bsd-2-clause
praekelt/molo.yourwords,praekelt/molo.yourwords
from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: - competitions = [] + competitions = YourWordsCompetition.objects.none() context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return []
Return None if there is no competition
## Code Before: from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: competitions = [] context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return [] ## Instruction: Return None if there is no competition ## Code After: from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: competitions = YourWordsCompetition.objects.none() context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return []
from django import template from copy import copy from molo.yourwords.models import (YourWordsCompetition, ThankYou, YourWordsCompetitionIndexPage) from molo.core.templatetags.core_tags import get_pages register = template.Library() @register.inclusion_tag( 'yourwords/your_words_competition_tag.html', takes_context=True ) def your_words_competition(context): context = copy(context) locale_code = context.get('locale_code') page = YourWordsCompetitionIndexPage.objects.live().all().first() if page: competitions = ( YourWordsCompetition.objects.child_of(page).filter( languages__language__is_main_language=True).specific()) else: - competitions = [] + competitions = YourWordsCompetition.objects.none() context.update({ 'competitions': get_pages(context, competitions, locale_code) }) return context @register.assignment_tag(takes_context=True) def load_thank_you_page_for_competition(context, competition): page = competition.get_main_language_page() locale = context.get('locale_code') qs = ThankYou.objects.child_of(page).filter( languages__language__is_main_language=True) if not locale: return qs if qs: return get_pages(context, qs, locale) else: return []
a42a6a54f732ca7eba700b867a3025739ad6a271
list_all_users_in_group.py
list_all_users_in_group.py
from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) if __name__ == "__main__": parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result))
from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) def main(): parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result)) if __name__ == "__main__": main()
Move main code to function because of pylint warning 'Invalid constant name'
Move main code to function because of pylint warning 'Invalid constant name'
Python
cc0-1.0
vazhnov/list_all_users_in_group
from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) - if __name__ == "__main__": + + def main(): parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result)) + + if __name__ == "__main__": + main() +
Move main code to function because of pylint warning 'Invalid constant name'
## Code Before: from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) if __name__ == "__main__": parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result)) ## Instruction: Move main code to function because of pylint warning 'Invalid constant name' ## Code After: from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) def main(): parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result)) if __name__ == "__main__": main()
from __future__ import print_function import grp import pwd import inspect import argparse def list_all_users_in_group(groupname): """Get list of all users of group. Get sorted list of all users of group GROUP, including users with main group GROUP. Origin in https://github.com/vazhnov/list_all_users_in_group """ try: group = grp.getgrnam(groupname) # On error "KeyError: 'getgrnam(): name not found: GROUP'" except KeyError: return None group_all_users_set = set(group.gr_mem) for user in pwd.getpwall(): if user.pw_gid == group.gr_gid: group_all_users_set.add(user.pw_name) return sorted(group_all_users_set) - if __name__ == "__main__": + + def main(): parser = argparse.ArgumentParser(description=inspect.getdoc(list_all_users_in_group), formatter_class=argparse.RawDescriptionHelpFormatter) parser.add_argument('-d', '--delimiter', default='\n', help='Use DELIMITER instead of newline for users delimiter') parser.add_argument('groupname', help='Group name') args = parser.parse_args() result = list_all_users_in_group(args.groupname) if result: print (args.delimiter.join(result)) + + + if __name__ == "__main__": + main()
4072f8ec6e1908d6e84859c8a0bd6c96562ea5cc
parts/plugins/x-shell.py
parts/plugins/x-shell.py
import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ])
import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, 'SNAPDIR=' + self.builddir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ]) # vim:set ts=4 noet:
Add "SNAPDIR" and simple vim modeline
Add "SNAPDIR" and simple vim modeline
Python
mit
infosiftr/snap-docker,docker-snap/docker,docker-snap/docker
import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, + 'SNAPDIR=' + self.builddir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ]) + # vim:set ts=4 noet: +
Add "SNAPDIR" and simple vim modeline
## Code Before: import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ]) ## Instruction: Add "SNAPDIR" and simple vim modeline ## Code After: import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, 'SNAPDIR=' + self.builddir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ]) # vim:set ts=4 noet:
import snapcraft class ShellPlugin(snapcraft.BasePlugin): @classmethod def schema(cls): schema = super().schema() schema['required'] = [] schema['properties']['shell'] = { 'type': 'string', 'default': '/bin/sh', } schema['required'].append('shell') schema['properties']['shell-flags'] = { 'type': 'array', 'items': { 'type': 'string', }, 'default': [], } schema['properties']['shell-command'] = { 'type': 'string', } schema['required'].append('shell-command') return schema def env(self, root): return super().env(root) + [ 'DESTDIR=' + self.installdir, + 'SNAPDIR=' + self.builddir, ] def build(self): super().build() return self.run([ self.options.shell, ] + self.options.shell_flags + [ '-c', self.options.shell_command, ]) + + # vim:set ts=4 noet:
2ad2d488b4d7b0997355c068646a6a38b2668dae
meetuppizza/tests.py
meetuppizza/tests.py
from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200)
from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) def test_page_contains_pizza(self): response = self.client.get('/') self.assertContains(response, "Pizza")
Add test that checks if landing page contains the word Pizza.
Add test that checks if landing page contains the word Pizza.
Python
mit
nicole-a-tesla/meetup.pizza,nicole-a-tesla/meetup.pizza
from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) + def test_page_contains_pizza(self): + response = self.client.get('/') + self.assertContains(response, "Pizza") +
Add test that checks if landing page contains the word Pizza.
## Code Before: from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) ## Instruction: Add test that checks if landing page contains the word Pizza. ## Code After: from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) def test_page_contains_pizza(self): response = self.client.get('/') self.assertContains(response, "Pizza")
from django.test import TestCase class Test(TestCase): def test_landing_page_is_there(self): response = self.client.get('/') self.assertEqual(response.status_code, 200) + + def test_page_contains_pizza(self): + response = self.client.get('/') + self.assertContains(response, "Pizza")
a35a25732159e4c8b5655755ce31ec4c3e6e7975
dummy_robot/dummy_robot_bringup/launch/dummy_robot_bringup.launch.py
dummy_robot/dummy_robot_bringup/launch/dummy_robot_bringup.launch.py
import os from ament_index_python.packages import get_package_share_directory from launch import LaunchDescription from launch_ros.actions import Node def generate_launch_description(): # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'), 'launch', 'single_rrbot.urdf') return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), Node(package='robot_state_publisher', node_executable='robot_state_publisher', output='screen', arguments=[urdf]), Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ])
import os from launch import LaunchDescription from launch_ros.actions import Node from launch_ros.substitutions import FindPackageShare def generate_launch_description(): pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup') urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf') with open(urdf_file, 'r') as infp: robot_desc = infp.read() rsp_params = {'robot_description': robot_desc} return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), Node(package='robot_state_publisher', node_executable='robot_state_publisher_node', output='screen', parameters=[rsp_params]), Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ])
Switch dummy_robot_bringup to use parameter for rsp.
Switch dummy_robot_bringup to use parameter for rsp. Signed-off-by: Chris Lalancette <281cd07d7578d97c83271fbbf2faddb83ab3791c@openrobotics.org>
Python
apache-2.0
ros2/demos,ros2/demos,ros2/demos,ros2/demos
import os - from ament_index_python.packages import get_package_share_directory from launch import LaunchDescription from launch_ros.actions import Node + from launch_ros.substitutions import FindPackageShare def generate_launch_description(): - # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch - urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'), - 'launch', 'single_rrbot.urdf') + pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup') + urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf') + with open(urdf_file, 'r') as infp: + robot_desc = infp.read() + rsp_params = {'robot_description': robot_desc} + return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), - Node(package='robot_state_publisher', node_executable='robot_state_publisher', + Node(package='robot_state_publisher', node_executable='robot_state_publisher_node', - output='screen', arguments=[urdf]), + output='screen', parameters=[rsp_params]), Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ])
Switch dummy_robot_bringup to use parameter for rsp.
## Code Before: import os from ament_index_python.packages import get_package_share_directory from launch import LaunchDescription from launch_ros.actions import Node def generate_launch_description(): # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'), 'launch', 'single_rrbot.urdf') return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), Node(package='robot_state_publisher', node_executable='robot_state_publisher', output='screen', arguments=[urdf]), Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ]) ## Instruction: Switch dummy_robot_bringup to use parameter for rsp. ## Code After: import os from launch import LaunchDescription from launch_ros.actions import Node from launch_ros.substitutions import FindPackageShare def generate_launch_description(): pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup') urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf') with open(urdf_file, 'r') as infp: robot_desc = infp.read() rsp_params = {'robot_description': robot_desc} return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), Node(package='robot_state_publisher', node_executable='robot_state_publisher_node', output='screen', parameters=[rsp_params]), Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ])
import os - from ament_index_python.packages import get_package_share_directory from launch import LaunchDescription from launch_ros.actions import Node + from launch_ros.substitutions import FindPackageShare def generate_launch_description(): - # TODO(wjwwood): Use a substitution to find share directory once this is implemented in launch - urdf = os.path.join(get_package_share_directory('dummy_robot_bringup'), - 'launch', 'single_rrbot.urdf') + pkg_share = FindPackageShare('dummy_robot_bringup').find('dummy_robot_bringup') + urdf_file = os.path.join(pkg_share, 'launch', 'single_rrbot.urdf') + with open(urdf_file, 'r') as infp: + robot_desc = infp.read() + rsp_params = {'robot_description': robot_desc} + return LaunchDescription([ Node(package='dummy_map_server', node_executable='dummy_map_server', output='screen'), - Node(package='robot_state_publisher', node_executable='robot_state_publisher', + Node(package='robot_state_publisher', node_executable='robot_state_publisher_node', ? +++++ - output='screen', arguments=[urdf]), ? ^^ - - ^^ + output='screen', parameters=[rsp_params]), ? + ^ ++ ^^^^^^^^^ Node(package='dummy_sensors', node_executable='dummy_joint_states', output='screen'), Node(package='dummy_sensors', node_executable='dummy_laser', output='screen') ])
456e5a63333e683b7167bf151b97a49a5cf5c5fe
app/models/job.py
app/models/job.py
from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION
from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
Rework the jod document model.
Rework the jod document model. * Add the created field that will store a datetime object. * Add reference to the kernel and the job inside the document, without relying on the Jod document name itself. Since we use the dash as a separator, and other job names can have dash in them, we cannot separate job from kernel easily.
Python
agpl-3.0
joyxu/kernelci-backend,joyxu/kernelci-backend,kernelci/kernelci-backend,kernelci/kernelci-backend,joyxu/kernelci-backend
from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' - def __init__(self, name): + def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) + + self._private = False + self._job = job + self._kernel = kernel + self._created = None @property def collection(self): return JOB_COLLECTION + @property + def private(self): + return self._private + + @private.setter + def private(self, value): + self._private = value + + @property + def job(self): + return self._job + + @job.setter + def job(self, value): + self._job = value + + @property + def kernel(self): + return self._kernel + + @kernel.setter + def kernel(self, value): + self._kernel = value + + @property + def created(self): + return self._created + + @created.setter + def created(self, value): + self._created = value + + def to_dict(self): + job_dict = super(JobDocument, self).to_dict() + job_dict['private'] = self._private + job_dict['job'] = self._job + job_dict['kernel'] = self._kernel + job_dict['created'] = str(self._created) + return job_dict +
Rework the jod document model.
## Code Before: from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name): super(JobDocument, self).__init__(name) @property def collection(self): return JOB_COLLECTION ## Instruction: Rework the jod document model. ## Code After: from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) self._private = False self._job = job self._kernel = kernel self._created = None @property def collection(self): return JOB_COLLECTION @property def private(self): return self._private @private.setter def private(self, value): self._private = value @property def job(self): return self._job @job.setter def job(self, value): self._job = value @property def kernel(self): return self._kernel @kernel.setter def kernel(self, value): self._kernel = value @property def created(self): return self._created @created.setter def created(self, value): self._created = value def to_dict(self): job_dict = super(JobDocument, self).to_dict() job_dict['private'] = self._private job_dict['job'] = self._job job_dict['kernel'] = self._kernel job_dict['created'] = str(self._created) return job_dict
from base import BaseDocument JOB_COLLECTION = 'job' class JobDocument(BaseDocument): JOB_ID_FORMAT = '%s-%s' - def __init__(self, name): + def __init__(self, name, job=None, kernel=None): super(JobDocument, self).__init__(name) + + self._private = False + self._job = job + self._kernel = kernel + self._created = None @property def collection(self): return JOB_COLLECTION + + @property + def private(self): + return self._private + + @private.setter + def private(self, value): + self._private = value + + @property + def job(self): + return self._job + + @job.setter + def job(self, value): + self._job = value + + @property + def kernel(self): + return self._kernel + + @kernel.setter + def kernel(self, value): + self._kernel = value + + @property + def created(self): + return self._created + + @created.setter + def created(self, value): + self._created = value + + def to_dict(self): + job_dict = super(JobDocument, self).to_dict() + job_dict['private'] = self._private + job_dict['job'] = self._job + job_dict['kernel'] = self._kernel + job_dict['created'] = str(self._created) + return job_dict
ab802204d84511765a701cad48e9e22dc4e84be1
tests/rules/conftest.py
tests/rules/conftest.py
import pytest from fmn.rules.cache import cache @pytest.fixture(autouse=True, scope="session") def configured_cache(): cache.configure()
import pytest from fmn.rules.cache import cache @pytest.fixture(autouse=True) def configured_cache(): if not cache.region.is_configured: cache.configure() yield cache.region.invalidate()
Fix intermittent failures of test_guard_http_exception
Fix intermittent failures of test_guard_http_exception Signed-off-by: Ryan Lerch <e809e25f3c554b2b195ccd768cd9a485288f896f@redhat.com>
Python
lgpl-2.1
fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn,fedora-infra/fmn
import pytest from fmn.rules.cache import cache - @pytest.fixture(autouse=True, scope="session") + @pytest.fixture(autouse=True) def configured_cache(): + if not cache.region.is_configured: - cache.configure() + cache.configure() + yield + cache.region.invalidate()
Fix intermittent failures of test_guard_http_exception
## Code Before: import pytest from fmn.rules.cache import cache @pytest.fixture(autouse=True, scope="session") def configured_cache(): cache.configure() ## Instruction: Fix intermittent failures of test_guard_http_exception ## Code After: import pytest from fmn.rules.cache import cache @pytest.fixture(autouse=True) def configured_cache(): if not cache.region.is_configured: cache.configure() yield cache.region.invalidate()
import pytest from fmn.rules.cache import cache - @pytest.fixture(autouse=True, scope="session") ? ----------------- + @pytest.fixture(autouse=True) def configured_cache(): + if not cache.region.is_configured: - cache.configure() + cache.configure() ? ++++ + yield + cache.region.invalidate()
cfb50f4ff62770c397634897e09497b74b396067
notifications/level_starting.py
notifications/level_starting.py
from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data
from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name data['message_data']['event_key'] = self.event.key_name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data
Add event key to comp level starting notification
Add event key to comp level starting notification
Python
mit
josephbisch/the-blue-alliance,synth3tk/the-blue-alliance,phil-lopreiato/the-blue-alliance,phil-lopreiato/the-blue-alliance,fangeugene/the-blue-alliance,bvisness/the-blue-alliance,nwalters512/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,the-blue-alliance/the-blue-alliance,verycumbersome/the-blue-alliance,bdaroz/the-blue-alliance,bvisness/the-blue-alliance,jaredhasenklein/the-blue-alliance,fangeugene/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,nwalters512/the-blue-alliance,the-blue-alliance/the-blue-alliance,jaredhasenklein/the-blue-alliance,phil-lopreiato/the-blue-alliance,the-blue-alliance/the-blue-alliance,the-blue-alliance/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,bdaroz/the-blue-alliance,verycumbersome/the-blue-alliance,jaredhasenklein/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,1fish2/the-blue-alliance,1fish2/the-blue-alliance,tsteward/the-blue-alliance,verycumbersome/the-blue-alliance,1fish2/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,bvisness/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,jaredhasenklein/the-blue-alliance,bdaroz/the-blue-alliance,tsteward/the-blue-alliance,tsteward/the-blue-alliance,nwalters512/the-blue-alliance,synth3tk/the-blue-alliance,synth3tk/the-blue-alliance,bdaroz/the-blue-alliance,synth3tk/the-blue-alliance,tsteward/the-blue-alliance,1fish2/the-blue-alliance,1fish2/the-blue-alliance,verycumbersome/the-blue-alliance,verycumbersome/the-blue-alliance,tsteward/the-blue-alliance,synth3tk/the-blue-alliance,nwalters512/the-blue-alliance,josephbisch/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,the-blue-alliance/the-blue-alliance,phil-lopreiato/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,1fish2/the-blue-alliance,the-blue-alliance/the-blue-alliance,synth3tk/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance
from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name + data['message_data']['event_key'] = self.event.key_name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data
Add event key to comp level starting notification
## Code Before: from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data ## Instruction: Add event key to comp level starting notification ## Code After: from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name data['message_data']['event_key'] = self.event.key_name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data
from consts.notification_type import NotificationType from notifications.base_notification import BaseNotification class CompLevelStartingNotification(BaseNotification): def __init__(self, match, event): self.match = match self.event = event def _build_dict(self): data = {} data['message_type'] = NotificationType.type_names[NotificationType.LEVEL_STARTING] data['message_data'] = {} data['message_data']['event_name'] = self.event.name + data['message_data']['event_key'] = self.event.key_name data['message_data']['comp_level'] = self.match.comp_level data['message_data']['scheduled_time'] = self.match.time return data
b4578d34adaa641dab5082f9d2bffe14c69649c5
detour/__init__.py
detour/__init__.py
from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover
from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover class DetourException(NotImplementedError): pass
Add a root exception for use if necessary.
Add a root exception for use if necessary.
Python
bsd-2-clause
kezabelle/wsgi-detour
from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover + + class DetourException(NotImplementedError): + pass +
Add a root exception for use if necessary.
## Code Before: from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover ## Instruction: Add a root exception for use if necessary. ## Code After: from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover class DetourException(NotImplementedError): pass
from __future__ import absolute_import from __future__ import unicode_literals __version_info__ = '0.1.0' __version__ = '0.1.0' version = '0.1.0' VERSION = '0.1.0' def get_version(): return version # pragma: no cover + + + class DetourException(NotImplementedError): + pass
55a1f6197800249b3ad13ec7c5358e907ea04c46
comics/comics/treadingground.py
comics/comics/treadingground.py
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): history_capable_days = 30 schedule = 'Mo,We,Fr' time_zone = -5 def crawl(self, pub_date): pass
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): schedule = None def crawl(self, pub_date): pass # Comic no longer published
Remove schedule for ended comic
Remove schedule for ended comic
Python
agpl-3.0
klette/comics,datagutten/comics,klette/comics,jodal/comics,datagutten/comics,klette/comics,jodal/comics,jodal/comics,jodal/comics,datagutten/comics,datagutten/comics
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): - history_capable_days = 30 - schedule = 'Mo,We,Fr' + schedule = None - time_zone = -5 def crawl(self, pub_date): - pass + pass # Comic no longer published
Remove schedule for ended comic
## Code Before: from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): history_capable_days = 30 schedule = 'Mo,We,Fr' time_zone = -5 def crawl(self, pub_date): pass ## Instruction: Remove schedule for ended comic ## Code After: from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): schedule = None def crawl(self, pub_date): pass # Comic no longer published
from comics.aggregator.crawler import CrawlerBase, CrawlerImage from comics.meta.base import MetaBase class Meta(MetaBase): name = 'Treading Ground' language = 'en' url = 'http://www.treadingground.com/' start_date = '2003-10-12' rights = 'Nick Wright' class Crawler(CrawlerBase): - history_capable_days = 30 - schedule = 'Mo,We,Fr' ? ^^ ^^ ---- + schedule = None ? ^ ^ - time_zone = -5 def crawl(self, pub_date): - pass + pass # Comic no longer published
1bc4507234d87b1ed246501165fa1d8138bf5ca6
cheddar/exceptions.py
cheddar/exceptions.py
class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): self.status_code = status_code
class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): super(NotFoundError, self).__init__() self.status_code = status_code
Fix for pypy compatibility: must super's __init__
Fix for pypy compatibility: must super's __init__
Python
apache-2.0
jessemyers/cheddar,jessemyers/cheddar
class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): + super(NotFoundError, self).__init__() self.status_code = status_code
Fix for pypy compatibility: must super's __init__
## Code Before: class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): self.status_code = status_code ## Instruction: Fix for pypy compatibility: must super's __init__ ## Code After: class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): super(NotFoundError, self).__init__() self.status_code = status_code
class BadRequestError(Exception): pass class ConflictError(Exception): pass class NotFoundError(Exception): def __init__(self, status_code=None): + super(NotFoundError, self).__init__() self.status_code = status_code
e6c43333c3939247534ddee4c419dcdcff5eda5f
spyder_terminal/server/rest/term_rest.py
spyder_terminal/server/rest/term_rest.py
"""Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" rows = int(self.get_argument('rows', None, 23)) cols = int(self.get_argument('cols', None, 73)) self.application.term_manager.resize_term(pid, rows, cols)
"""Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) self.application.term_manager.resize_term(pid, rows, cols)
Change default terminal resize arguments
Change default terminal resize arguments
Python
mit
andfoy/spyder-terminal,spyder-ide/spyder-terminal,spyder-ide/spyder-terminal,andfoy/spyder-terminal,andfoy/spyder-terminal,spyder-ide/spyder-terminal,spyder-ide/spyder-terminal
"""Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" - rows = int(self.get_argument('rows', None, 23)) + rows = int(self.get_argument('rows', default=23)) - cols = int(self.get_argument('cols', None, 73)) + cols = int(self.get_argument('cols', default=73)) self.application.term_manager.resize_term(pid, rows, cols)
Change default terminal resize arguments
## Code Before: """Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" rows = int(self.get_argument('rows', None, 23)) cols = int(self.get_argument('cols', None, 73)) self.application.term_manager.resize_term(pid, rows, cols) ## Instruction: Change default terminal resize arguments ## Code After: """Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) self.application.term_manager.resize_term(pid, rows, cols)
"""Main HTTP routes request handlers.""" import tornado.web import tornado.escape from os import getcwd class MainHandler(tornado.web.RequestHandler): """Handles creation of new terminals.""" @tornado.gen.coroutine def post(self): """POST verb: Create a new terminal.""" rows = int(self.get_argument('rows', default=23)) cols = int(self.get_argument('cols', default=73)) cwd = self.get_cookie('cwd', default=getcwd()) self.application.logger.info('CWD: {0}'.format(cwd)) self.application.logger.info('Size: ({0}, {1})'.format(cols, rows)) pid = yield self.application.term_manager.create_term(rows, cols, cwd) self.write(pid) class ResizeHandler(tornado.web.RequestHandler): """Handles resizing of terminals.""" @tornado.gen.coroutine def post(self, pid): """POST verb: Resize a terminal.""" - rows = int(self.get_argument('rows', None, 23)) ? ^^^ ^^ + rows = int(self.get_argument('rows', default=23)) ? ^ ^^^^^^ - cols = int(self.get_argument('cols', None, 73)) ? ^^^ ^^ + cols = int(self.get_argument('cols', default=73)) ? ^ ^^^^^^ self.application.term_manager.resize_term(pid, rows, cols)
746510dc0b939fe11a2b025805678a0829cf814a
handler/minion_server.py
handler/minion_server.py
import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main()
import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): """Start a worker. Message format: { 'image': 'image name' 'numprocs': number of workers, 'args': 'extra arguments for "docker run -d image ..."' } """ supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main()
Document message format for minion server
Document message format for minion server
Python
mit
waltermoreira/adama-minion
import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): + """Start a worker. + + Message format: + + { + 'image': 'image name' + 'numprocs': number of workers, + 'args': 'extra arguments for "docker run -d image ..."' + } + + """ supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main()
Document message format for minion server
## Code Before: import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main() ## Instruction: Document message format for minion server ## Code After: import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): """Start a worker. Message format: { 'image': 'image name' 'numprocs': number of workers, 'args': 'extra arguments for "docker run -d image ..."' } """ supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main()
import server import supervisor class MinionServer(server.Server): def __init__(self, ip, port): super(MinionServer, self).__init__(ip, port) def handle(self, data): + """Start a worker. + + Message format: + + { + 'image': 'image name' + 'numprocs': number of workers, + 'args': 'extra arguments for "docker run -d image ..."' + } + + """ supervisor.start( 'worker.conf', target='worker_{}'.format(data['image']), image=data['image'], numprocs=data.get('numprocs', 1), args=data.get('args', '')) return {'status': 'ok'} def main(): server = MinionServer('*', 1234) server.start() server.join() if __name__ == '__main__': main()
414dd0b03b3e4eabc11f848f79d681f3a284380e
pygcvs/helpers.py
pygcvs/helpers.py
from .parser import GcvsParser try: import ephem except ImportError: ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ if ephem is None: raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body
from .parser import GcvsParser try: import ephem except ImportError: # pragma: no cover ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ if ephem is None: # pragma: no cover raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body
Exclude missing ephem from coverage
Exclude missing ephem from coverage
Python
mit
zsiciarz/pygcvs
from .parser import GcvsParser try: import ephem - except ImportError: + except ImportError: # pragma: no cover ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ - if ephem is None: + if ephem is None: # pragma: no cover raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body
Exclude missing ephem from coverage
## Code Before: from .parser import GcvsParser try: import ephem except ImportError: ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ if ephem is None: raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body ## Instruction: Exclude missing ephem from coverage ## Code After: from .parser import GcvsParser try: import ephem except ImportError: # pragma: no cover ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ if ephem is None: # pragma: no cover raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body
from .parser import GcvsParser try: import ephem - except ImportError: + except ImportError: # pragma: no cover ephem = None def read_gcvs(filename): """ Reads variable star data in `GCVS format`_. :param filename: path to GCVS data file (usually ``iii.dat``) .. _`GCVS format`: http://www.sai.msu.su/gcvs/gcvs/iii/html/ """ with open(filename, 'r') as fp: parser = GcvsParser(fp) for star in parser: yield star def dict_to_body(star_dict): """ Converts a dictionary of variable star data to a `Body` instance. Requires `PyEphem <http://rhodesmill.org/pyephem/>`_ to be installed. """ - if ephem is None: + if ephem is None: # pragma: no cover raise NotImplementedError("Please install PyEphem in order to use dict_to_body.") body = ephem.FixedBody() body.name = star_dict['name'] body._ra = ephem.hours(str(star_dict['ra'])) body._dec = ephem.degrees(str(star_dict['dec'])) body._epoch = ephem.J2000 return body
4949b1051656566ce544a8240b0328a61259868a
migrations/versions/139_add_ns_index_to_contact_and_event.py
migrations/versions/139_add_ns_index_to_contact_and_event.py
# revision identifiers, used by Alembic. revision = '1fd7b3e0b662' down_revision = '2d8a350b4885' from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.")
# revision identifiers, used by Alembic. revision = '1fd7b3e0b662' down_revision = '5305d4ae30b4' from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.")
Fix migration history bug introduced with merge
Fix migration history bug introduced with merge
Python
agpl-3.0
wakermahmud/sync-engine,EthanBlackburn/sync-engine,Eagles2F/sync-engine,EthanBlackburn/sync-engine,gale320/sync-engine,wakermahmud/sync-engine,nylas/sync-engine,jobscore/sync-engine,nylas/sync-engine,jobscore/sync-engine,closeio/nylas,gale320/sync-engine,wakermahmud/sync-engine,closeio/nylas,PriviPK/privipk-sync-engine,PriviPK/privipk-sync-engine,jobscore/sync-engine,nylas/sync-engine,closeio/nylas,PriviPK/privipk-sync-engine,jobscore/sync-engine,ErinCall/sync-engine,Eagles2F/sync-engine,PriviPK/privipk-sync-engine,nylas/sync-engine,Eagles2F/sync-engine,EthanBlackburn/sync-engine,ErinCall/sync-engine,wakermahmud/sync-engine,wakermahmud/sync-engine,Eagles2F/sync-engine,ErinCall/sync-engine,closeio/nylas,gale320/sync-engine,PriviPK/privipk-sync-engine,gale320/sync-engine,Eagles2F/sync-engine,ErinCall/sync-engine,EthanBlackburn/sync-engine,EthanBlackburn/sync-engine,ErinCall/sync-engine,gale320/sync-engine
# revision identifiers, used by Alembic. revision = '1fd7b3e0b662' - down_revision = '2d8a350b4885' + down_revision = '5305d4ae30b4' from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.")
Fix migration history bug introduced with merge
## Code Before: # revision identifiers, used by Alembic. revision = '1fd7b3e0b662' down_revision = '2d8a350b4885' from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.") ## Instruction: Fix migration history bug introduced with merge ## Code After: # revision identifiers, used by Alembic. revision = '1fd7b3e0b662' down_revision = '5305d4ae30b4' from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.")
# revision identifiers, used by Alembic. revision = '1fd7b3e0b662' - down_revision = '2d8a350b4885' ? ^ ^ - --- + down_revision = '5305d4ae30b4' ? ^^^^ ^ + from alembic import op def upgrade(): op.create_index( 'ix_contact_ns_uid_provider_name', 'contact', ['namespace_id', 'uid', 'provider_name'], unique=False) op.create_index( 'ix_event_ns_uid_provider_name', 'event', ['namespace_id', 'uid', 'provider_name'], unique=False) def downgrade(): raise Exception("Don't bother.")
ab8fc00a7dc6618d23e06f06e125da5ee69b2dba
event_registration_hr_contract/__openerp__.py
event_registration_hr_contract/__openerp__.py
{ "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', 'hr_contract_stages' ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, }
{ "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, }
Remove the dependence with the module hr_contract_stage.
[IMP] event_registration_hr_contract: Remove the dependence with the module hr_contract_stage.
Python
agpl-3.0
avanzosc/event-wip
{ "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', - 'hr_contract_stages' ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, }
Remove the dependence with the module hr_contract_stage.
## Code Before: { "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', 'hr_contract_stages' ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, } ## Instruction: Remove the dependence with the module hr_contract_stage. ## Code After: { "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, }
{ "name": "Event Registration Hr Contract", 'version': '8.0.1.1.0', 'license': "AGPL-3", 'author': "AvanzOSC", 'website': "http://www.avanzosc.es", 'contributors': [ "Ana Juaristi <anajuaristi@avanzosc.es>", "Alfredo de la Fuente <alfredodelafuente@avanzosc.es", ], "category": "Event Management", "depends": [ 'event_track_presence_hr_holidays', - 'hr_contract_stages' ], "data": [ 'wizard/wiz_calculate_employee_calendar_view.xml', 'wizard/wiz_event_append_assistant_view.xml', 'views/event_event_view.xml', 'views/hr_contract_view.xml', 'views/event_track_presence_view.xml', 'views/res_partner_calendar_view.xml', 'views/res_partner_calendar_day_view.xml' ], "installable": True, }
6795e02c14fa99da2c0812fe6694bbd503f89ad1
tests/mock_vws/test_invalid_given_id.py
tests/mock_vws/test_invalid_given_id.py
import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, endpoint: TargetAPIEndpoint, target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, )
import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, any_endpoint: TargetAPIEndpoint, target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ endpoint = any_endpoint if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, )
Use any_endpoint on invalid id test
Use any_endpoint on invalid id test
Python
mit
adamtheturtle/vws-python,adamtheturtle/vws-python
import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, - endpoint: TargetAPIEndpoint, + any_endpoint: TargetAPIEndpoint, target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ + endpoint = any_endpoint if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, )
Use any_endpoint on invalid id test
## Code Before: import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, endpoint: TargetAPIEndpoint, target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, ) ## Instruction: Use any_endpoint on invalid id test ## Code After: import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, any_endpoint: TargetAPIEndpoint, target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ endpoint = any_endpoint if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, )
import pytest import requests from requests import codes from mock_vws._constants import ResultCodes from tests.mock_vws.utils import ( TargetAPIEndpoint, VuforiaDatabaseKeys, assert_vws_failure, delete_target, ) @pytest.mark.usefixtures('verify_mock_vuforia') class TestInvalidGivenID: """ Tests for giving an invalid ID to endpoints which require a target ID to be given. """ def test_not_real_id( self, vuforia_database_keys: VuforiaDatabaseKeys, - endpoint: TargetAPIEndpoint, + any_endpoint: TargetAPIEndpoint, ? ++++ target_id: str, ) -> None: """ A `NOT_FOUND` error is returned when an endpoint is given a target ID of a target which does not exist. """ + endpoint = any_endpoint if not endpoint.prepared_request.path_url.endswith(target_id): return delete_target( vuforia_database_keys=vuforia_database_keys, target_id=target_id, ) session = requests.Session() response = session.send( # type: ignore request=endpoint.prepared_request, ) assert_vws_failure( response=response, status_code=codes.NOT_FOUND, result_code=ResultCodes.UNKNOWN_TARGET, )
de42731ab97a7d4272c44cc750891906aa5b4417
buildlet/runner/ipythonparallel.py
buildlet/runner/ipythonparallel.py
import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): self.view.wait(self.results.values())
import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): for r in self.results.values(): r.get()
Raise error if any in IPythonParallelRunner.wait_tasks
Raise error if any in IPythonParallelRunner.wait_tasks
Python
bsd-3-clause
tkf/buildlet
import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): - self.view.wait(self.results.values()) + for r in self.results.values(): + r.get()
Raise error if any in IPythonParallelRunner.wait_tasks
## Code Before: import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): self.view.wait(self.results.values()) ## Instruction: Raise error if any in IPythonParallelRunner.wait_tasks ## Code After: import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): for r in self.results.values(): r.get()
import IPython.parallel from .simple import SimpleRunner from .mixinparallel import MixInParallelRunner class IPythonParallelRunner(MixInParallelRunner, SimpleRunner): def submit_tasks(self): self.client = IPython.parallel.Client() self.view = view = self.client.load_balanced_view() self.results = results = {} for node in self.sorted_nodes(): deps = [results[n] for n in self.graph.predecessors(node)] with view.temp_flags(after=deps): results[node] = view.apply_async(self.run_func, self.nodetaskmap[node]) def wait_tasks(self): - self.view.wait(self.results.values()) + for r in self.results.values(): + r.get()
d565fdab9cefc080ff3127f036c19e95cba73f6e
tests/test_udacity.py
tests/test_udacity.py
import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) def tearDown(self): pass if __name__ == '__main__': unittest.main()
import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) def test_udacity_api_mongofy_courses(self): course = self.udacity_test_object.mongofy_courses()[0] self.assertEqual(course['title'], 'Intro to Computer Science') def tearDown(self): pass if __name__ == '__main__': unittest.main()
Add unit test for mongofy_courses of udacity module
Add unit test for mongofy_courses of udacity module
Python
mit
ueg1990/mooc_aggregator_restful_api
import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) + def test_udacity_api_mongofy_courses(self): + course = self.udacity_test_object.mongofy_courses()[0] + self.assertEqual(course['title'], 'Intro to Computer Science') + def tearDown(self): pass if __name__ == '__main__': unittest.main()
Add unit test for mongofy_courses of udacity module
## Code Before: import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) def tearDown(self): pass if __name__ == '__main__': unittest.main() ## Instruction: Add unit test for mongofy_courses of udacity module ## Code After: import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) def test_udacity_api_mongofy_courses(self): course = self.udacity_test_object.mongofy_courses()[0] self.assertEqual(course['title'], 'Intro to Computer Science') def tearDown(self): pass if __name__ == '__main__': unittest.main()
import unittest from mooc_aggregator_restful_api import udacity class UdacityTestCase(unittest.TestCase): ''' Unit Tests for module udacity ''' def setUp(self): self.udacity_test_object = udacity.UdacityAPI() def test_udacity_api_response(self): self.assertEqual(self.udacity_test_object.status_code(), 200) + def test_udacity_api_mongofy_courses(self): + course = self.udacity_test_object.mongofy_courses()[0] + self.assertEqual(course['title'], 'Intro to Computer Science') + def tearDown(self): pass if __name__ == '__main__': unittest.main()
d7f43a15a2e4535728e7ec5d3cb550af3eed1590
h2o-py/h2o/tree/__init__.py
h2o-py/h2o/tree/__init__.py
from .tree import H2OTree from .tree import H2ONode __all__ = ["H2OTree", "H2ONode"]
from .tree import H2OTree from .tree import H2ONode from .tree import H2OSplitNode from .tree import H2OLeafNode __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
Include H2OSplitNode & H2OLeafNode in __all__
Include H2OSplitNode & H2OLeafNode in __all__
Python
apache-2.0
michalkurka/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,h2oai/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,michalkurka/h2o-3,michalkurka/h2o-3,h2oai/h2o-3
from .tree import H2OTree from .tree import H2ONode + from .tree import H2OSplitNode + from .tree import H2OLeafNode - __all__ = ["H2OTree", "H2ONode"] + __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"] +
Include H2OSplitNode & H2OLeafNode in __all__
## Code Before: from .tree import H2OTree from .tree import H2ONode __all__ = ["H2OTree", "H2ONode"] ## Instruction: Include H2OSplitNode & H2OLeafNode in __all__ ## Code After: from .tree import H2OTree from .tree import H2ONode from .tree import H2OSplitNode from .tree import H2OLeafNode __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
from .tree import H2OTree from .tree import H2ONode + from .tree import H2OSplitNode + from .tree import H2OLeafNode - __all__ = ["H2OTree", "H2ONode"] + __all__ = ["H2OTree", "H2ONode", "H2OSplitNode", "H2OLeafNode"]
7e6dc283dbecf4bf9674559198b4a2c06e9f4c2e
spacy/tests/regression/test_issue1799.py
spacy/tests/regression/test_issue1799.py
'''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1
'''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' from __future__ import unicode_literals import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1
Fix unicode import in test
Fix unicode import in test
Python
mit
aikramer2/spaCy,aikramer2/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,explosion/spaCy,spacy-io/spaCy,recognai/spaCy,honnibal/spaCy,spacy-io/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,aikramer2/spaCy,spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,recognai/spaCy,explosion/spaCy,recognai/spaCy,honnibal/spaCy,spacy-io/spaCy
'''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' + from __future__ import unicode_literals import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1
Fix unicode import in test
## Code Before: '''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1 ## Instruction: Fix unicode import in test ## Code After: '''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' from __future__ import unicode_literals import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1
'''Test sentence boundaries are deserialized correctly, even for non-projective sentences.''' + from __future__ import unicode_literals import pytest import numpy from ... tokens import Doc from ... vocab import Vocab from ... attrs import HEAD, DEP def test_issue1799(): problem_sentence = 'Just what I was looking for.' heads_deps = numpy.asarray([[1, 397], [4, 436], [2, 426], [1, 402], [0, 8206900633647566924], [18446744073709551615, 440], [18446744073709551614, 442]], dtype='uint64') doc = Doc(Vocab(), words='Just what I was looking for .'.split()) doc.vocab.strings.add('ROOT') doc = doc.from_array([HEAD, DEP], heads_deps) assert len(list(doc.sents)) == 1
8a080a94300403487dce023eec8467832af8ae79
tests/core/migrations/0004_bookwithchapters.py
tests/core/migrations/0004_bookwithchapters.py
from __future__ import unicode_literals from django import VERSION from django.db import migrations, models if VERSION >= (1, 8): from django.contrib.postgres.fields import ArrayField chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) else: chapters_field = models.Field() # Dummy field class PostgresOnlyCreateModel(migrations.CreateModel): def database_forwards(self, app_label, schema_editor, from_state, to_state): if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state) def database_backwards(self, app_label, schema_editor, from_state, to_state): if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state) class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] operations = [ PostgresOnlyCreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ]
from __future__ import unicode_literals from django import VERSION from django.db import migrations, models can_use_arrayfield = False chapters_field = models.Field() # Dummy field if VERSION >= (1, 8): try: from django.contrib.postgres.fields import ArrayField chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) can_use_arrayfield = True except ImportError: # We can't use ArrayField if psycopg2 is not installed pass class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] operations = [] pg_only_operations = [ migrations.CreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ] def apply(self, project_state, schema_editor, collect_sql=False): if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"): self.operations = self.operations + self.pg_only_operations return super(Migration, self).apply(project_state, schema_editor, collect_sql)
Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed
Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed
Python
bsd-2-clause
brillgen/django-import-export,copperleaftech/django-import-export,bmihelac/django-import-export,jnns/django-import-export,copperleaftech/django-import-export,jnns/django-import-export,brillgen/django-import-export,PetrDlouhy/django-import-export,jnns/django-import-export,PetrDlouhy/django-import-export,brillgen/django-import-export,django-import-export/django-import-export,bmihelac/django-import-export,django-import-export/django-import-export,PetrDlouhy/django-import-export,copperleaftech/django-import-export,daniell/django-import-export,bmihelac/django-import-export,PetrDlouhy/django-import-export,daniell/django-import-export,copperleaftech/django-import-export,bmihelac/django-import-export,jnns/django-import-export,brillgen/django-import-export,django-import-export/django-import-export,daniell/django-import-export,django-import-export/django-import-export,daniell/django-import-export
from __future__ import unicode_literals from django import VERSION from django.db import migrations, models + + can_use_arrayfield = False + chapters_field = models.Field() # Dummy field if VERSION >= (1, 8): + try: - from django.contrib.postgres.fields import ArrayField + from django.contrib.postgres.fields import ArrayField - chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) + chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) + can_use_arrayfield = True + except ImportError: + # We can't use ArrayField if psycopg2 is not installed + pass - else: - chapters_field = models.Field() # Dummy field - - - class PostgresOnlyCreateModel(migrations.CreateModel): - def database_forwards(self, app_label, schema_editor, from_state, to_state): - if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): - super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state) - - def database_backwards(self, app_label, schema_editor, from_state, to_state): - if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): - super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state) class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] - operations = [ + operations = [] - PostgresOnlyCreateModel( + + pg_only_operations = [ + migrations.CreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ] + def apply(self, project_state, schema_editor, collect_sql=False): + if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"): + self.operations = self.operations + self.pg_only_operations + return super(Migration, self).apply(project_state, schema_editor, collect_sql) +
Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed
## Code Before: from __future__ import unicode_literals from django import VERSION from django.db import migrations, models if VERSION >= (1, 8): from django.contrib.postgres.fields import ArrayField chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) else: chapters_field = models.Field() # Dummy field class PostgresOnlyCreateModel(migrations.CreateModel): def database_forwards(self, app_label, schema_editor, from_state, to_state): if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state) def database_backwards(self, app_label, schema_editor, from_state, to_state): if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state) class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] operations = [ PostgresOnlyCreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ] ## Instruction: Refactor migration script to handle case where Postgres is being used but psycopg2 isn't installed ## Code After: from __future__ import unicode_literals from django import VERSION from django.db import migrations, models can_use_arrayfield = False chapters_field = models.Field() # Dummy field if VERSION >= (1, 8): try: from django.contrib.postgres.fields import ArrayField chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) can_use_arrayfield = True except ImportError: # We can't use ArrayField if psycopg2 is not installed pass class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] operations = [] pg_only_operations = [ migrations.CreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ] def apply(self, project_state, schema_editor, collect_sql=False): if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"): self.operations = self.operations + self.pg_only_operations return super(Migration, self).apply(project_state, schema_editor, collect_sql)
from __future__ import unicode_literals from django import VERSION from django.db import migrations, models + + can_use_arrayfield = False + chapters_field = models.Field() # Dummy field if VERSION >= (1, 8): + try: - from django.contrib.postgres.fields import ArrayField + from django.contrib.postgres.fields import ArrayField ? ++++ - chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) + chapters_field = ArrayField(base_field=models.CharField(max_length=100), default=list, size=None) ? ++++ + can_use_arrayfield = True + except ImportError: + # We can't use ArrayField if psycopg2 is not installed + pass - else: - chapters_field = models.Field() # Dummy field - - - class PostgresOnlyCreateModel(migrations.CreateModel): - def database_forwards(self, app_label, schema_editor, from_state, to_state): - if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): - super(PostgresOnlyCreateModel, self).database_forwards(app_label, schema_editor, from_state, to_state) - - def database_backwards(self, app_label, schema_editor, from_state, to_state): - if VERSION >= (1, 8) and schema_editor.connection.vendor.startswith("postgres"): - super(PostgresOnlyCreateModel, self).database_backwards(app_label, schema_editor, from_state, to_state) class Migration(migrations.Migration): dependencies = [ ('core', '0003_withfloatfield'), ] - operations = [ + operations = [] ? + - PostgresOnlyCreateModel( + + pg_only_operations = [ + migrations.CreateModel( name='BookWithChapters', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=100, verbose_name='Book name')), ('chapters', chapters_field) ], ), ] + + def apply(self, project_state, schema_editor, collect_sql=False): + if can_use_arrayfield and schema_editor.connection.vendor.startswith("postgres"): + self.operations = self.operations + self.pg_only_operations + return super(Migration, self).apply(project_state, schema_editor, collect_sql)
71088ebbed3f6060def0455814036185c70ba194
shopify_auth/context_processors.py
shopify_auth/context_processors.py
import shopify def current_shop(request): if not shopify.ShopifyResource.site: return {'current_shop': None} return {'current_shop': shopify.Shop.current()}
from django.conf import settings import shopify def shopify_context(request): return { 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None, 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY, }
Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information.
Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information.
Python
mit
funkybob/django-shopify-auth,RafaAguilar/django-shopify-auth,discolabs/django-shopify-auth,RafaAguilar/django-shopify-auth,discolabs/django-shopify-auth,funkybob/django-shopify-auth
+ from django.conf import settings import shopify - def current_shop(request): - if not shopify.ShopifyResource.site: - return {'current_shop': None} - return {'current_shop': shopify.Shop.current()} + def shopify_context(request): + return { + 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None, + 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY, + }
Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information.
## Code Before: import shopify def current_shop(request): if not shopify.ShopifyResource.site: return {'current_shop': None} return {'current_shop': shopify.Shop.current()} ## Instruction: Rename `current_shop` context processor to `shopify_context`, and add a little more useful Shopify information. ## Code After: from django.conf import settings import shopify def shopify_context(request): return { 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None, 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY, }
+ from django.conf import settings import shopify - def current_shop(request): - if not shopify.ShopifyResource.site: - return {'current_shop': None} - return {'current_shop': shopify.Shop.current()} + def shopify_context(request): + return { + 'shopify_current_shop': shopify.Shop.current() if shopify.ShopifyResource.site else None, + 'shopify_app_api_key': settings.SHOPIFY_APP_API_KEY, + }
700fa0144c5276d8e31c01a243340f6cbac07e8f
sentry/client/handlers.py
sentry/client/handlers.py
import logging class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client get_client().create_from_record(record)
import logging import sys class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client # Avoid typical config issues by overriding loggers behavior if record.name == 'sentry.errors': print >> sys.stderr, record.message return get_client().create_from_record(record)
Add a safety net for recursive logging
Add a safety net for recursive logging
Python
bsd-3-clause
ewdurbin/sentry,hongliang5623/sentry,Photonomie/raven-python,ewdurbin/raven-python,daevaorn/sentry,NickPresta/sentry,dcramer/sentry-old,looker/sentry,NickPresta/sentry,chayapan/django-sentry,Kryz/sentry,BuildingLink/sentry,gencer/sentry,boneyao/sentry,danriti/raven-python,imankulov/sentry,nikolas/raven-python,mvaled/sentry,gencer/sentry,inspirehep/raven-python,arthurlogilab/raven-python,daevaorn/sentry,recht/raven-python,icereval/raven-python,drcapulet/sentry,ewdurbin/sentry,akheron/raven-python,boneyao/sentry,JamesMura/sentry,zenefits/sentry,johansteffner/raven-python,ewdurbin/sentry,gg7/sentry,SilentCircle/sentry,dbravender/raven-python,alex/raven,alex/sentry,zenefits/sentry,someonehan/raven-python,JTCunning/sentry,JTCunning/sentry,jbarbuto/raven-python,beniwohli/apm-agent-python,vperron/sentry,someonehan/raven-python,korealerts1/sentry,arthurlogilab/raven-python,Kronuz/django-sentry,mvaled/sentry,beni55/sentry,llonchj/sentry,mitsuhiko/raven,beni55/sentry,tarkatronic/opbeat_python,BuildingLink/sentry,WoLpH/django-sentry,jbarbuto/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,nikolas/raven-python,gencer/sentry,recht/raven-python,alexm92/sentry,jbarbuto/raven-python,wong2/sentry,pauloschilling/sentry,lepture/raven-python,felixbuenemann/sentry,nicholasserra/sentry,icereval/raven-python,percipient/raven-python,jmagnusson/raven-python,Kronuz/django-sentry,Natim/sentry,icereval/raven-python,SilentCircle/sentry,songyi199111/sentry,argonemyth/sentry,beeftornado/sentry,camilonova/sentry,NickPresta/sentry,smarkets/raven-python,akalipetis/raven-python,recht/raven-python,chayapan/django-sentry,inspirehep/raven-python,patrys/opbeat_python,SilentCircle/sentry,zenefits/sentry,ronaldevers/raven-python,beniwohli/apm-agent-python,kevinlondon/sentry,jean/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,mvaled/sentry,BayanGroup/sentry,gencer/sentry,jmp0xf/raven-python,kevinastone/sentry,openlabs/raven,daevaorn/sentry,patrys/opbeat_python,wong2/sentry,JamesMura/sentry,WoLpH/django-sentry,Photonomie/raven-python,nikolas/raven-python,Kryz/sentry,looker/sentry,songyi199111/sentry,ifduyue/sentry,NickPresta/sentry,1tush/sentry,mitsuhiko/sentry,fotinakis/sentry,tarkatronic/opbeat_python,dirtycoder/opbeat_python,mvaled/sentry,vperron/sentry,korealerts1/sentry,dirtycoder/opbeat_python,jean/sentry,nikolas/raven-python,alex/sentry,jmagnusson/raven-python,percipient/raven-python,argonemyth/sentry,hongliang5623/sentry,JackDanger/sentry,kevinlondon/sentry,pauloschilling/sentry,fuziontech/sentry,tbarbugli/sentry_fork,rdio/sentry,smarkets/raven-python,looker/sentry,BuildingLink/sentry,camilonova/sentry,JTCunning/sentry,ifduyue/sentry,someonehan/raven-python,jean/sentry,Kryz/sentry,hzy/raven-python,arthurlogilab/raven-python,JamesMura/sentry,felixbuenemann/sentry,ronaldevers/raven-python,fuziontech/sentry,ifduyue/sentry,llonchj/sentry,SilentCircle/sentry,1tush/sentry,kevinastone/sentry,looker/sentry,zenefits/sentry,pauloschilling/sentry,alex/sentry,jbarbuto/raven-python,looker/sentry,collective/mr.poe,gg7/sentry,felixbuenemann/sentry,1tush/sentry,jean/sentry,ronaldevers/raven-python,Natim/sentry,BuildingLink/sentry,daikeren/opbeat_python,Natim/sentry,Goldmund-Wyldebeast-Wunderliebe/raven-python,wujuguang/sentry,wujuguang/sentry,akalipetis/raven-python,JackDanger/sentry,BayanGroup/sentry,arthurlogilab/raven-python,boneyao/sentry,BayanGroup/sentry,mitsuhiko/raven,tbarbugli/sentry_fork,dcramer/sentry-old,Goldmund-Wyldebeast-Wunderliebe/raven-python,daikeren/opbeat_python,ticosax/opbeat_python,beeftornado/sentry,jmp0xf/raven-python,dbravender/raven-python,jokey2k/sentry,drcapulet/sentry,WoLpH/django-sentry,ticosax/opbeat_python,fotinakis/sentry,primepix/django-sentry,dcramer/sentry-old,BuildingLink/sentry,ewdurbin/raven-python,mvaled/sentry,ngonzalvez/sentry,llonchj/sentry,beni55/sentry,smarkets/raven-python,TedaLIEz/sentry,lepture/raven-python,hongliang5623/sentry,primepix/django-sentry,hzy/raven-python,akheron/raven-python,lopter/raven-python-old,Kronuz/django-sentry,nicholasserra/sentry,daevaorn/sentry,TedaLIEz/sentry,fotinakis/sentry,rdio/sentry,hzy/raven-python,vperron/sentry,imankulov/sentry,ifduyue/sentry,mvaled/sentry,korealerts1/sentry,ngonzalvez/sentry,tbarbugli/sentry_fork,rdio/sentry,getsentry/raven-python,chayapan/django-sentry,rdio/sentry,johansteffner/raven-python,getsentry/raven-python,jean/sentry,TedaLIEz/sentry,patrys/opbeat_python,ngonzalvez/sentry,camilonova/sentry,JackDanger/sentry,smarkets/raven-python,akheron/raven-python,kevinastone/sentry,Photonomie/raven-python,JamesMura/sentry,wujuguang/sentry,ticosax/opbeat_python,inspirehep/raven-python,beeftornado/sentry,argonemyth/sentry,beniwohli/apm-agent-python,tarkatronic/opbeat_python,ewdurbin/raven-python,fuziontech/sentry,zenefits/sentry,songyi199111/sentry,jokey2k/sentry,daikeren/opbeat_python,gencer/sentry,patrys/opbeat_python,dirtycoder/opbeat_python,icereval/raven-python,getsentry/raven-python,jmp0xf/raven-python,danriti/raven-python,nicholasserra/sentry,inspirehep/raven-python,wong2/sentry,gg7/sentry,imankulov/sentry,percipient/raven-python,johansteffner/raven-python,ifduyue/sentry,beniwohli/apm-agent-python,danriti/raven-python,alexm92/sentry,fotinakis/sentry,lepture/raven-python,akalipetis/raven-python,primepix/django-sentry,jmagnusson/raven-python,jokey2k/sentry,alexm92/sentry,JamesMura/sentry,drcapulet/sentry,kevinlondon/sentry,dbravender/raven-python,mitsuhiko/sentry
import logging + import sys class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client + # Avoid typical config issues by overriding loggers behavior + if record.name == 'sentry.errors': + print >> sys.stderr, record.message + return + get_client().create_from_record(record)
Add a safety net for recursive logging
## Code Before: import logging class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client get_client().create_from_record(record) ## Instruction: Add a safety net for recursive logging ## Code After: import logging import sys class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client # Avoid typical config issues by overriding loggers behavior if record.name == 'sentry.errors': print >> sys.stderr, record.message return get_client().create_from_record(record)
import logging + import sys class SentryHandler(logging.Handler): def emit(self, record): from sentry.client.models import get_client + # Avoid typical config issues by overriding loggers behavior + if record.name == 'sentry.errors': + print >> sys.stderr, record.message + return + get_client().create_from_record(record)
f4f439f24dceb0c68f05a90196b3e4b525d1aa7a
setup.py
setup.py
import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', )
import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: DFSG approved', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Indexing/Search', 'Topic :: Software Development :: Libraries'], )
Add some trove classifiers to the package metadata
Add some trove classifiers to the package metadata
Python
mit
rlskoeser/sunburnt,pixbuffer/sunburnt-spatial,anmar/sunburnt,rlskoeser/sunburnt,pixbuffer/sunburnt-spatial,qmssof/sunburnt,tow/sunburnt,anmar/sunburnt
import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', + classifiers=[ + 'Development Status :: 3 - Alpha', + 'Intended Audience :: Developers', + 'License :: DFSG approved', + 'Programming Language :: Python', + 'Topic :: Internet :: WWW/HTTP :: Indexing/Search', + 'Topic :: Software Development :: Libraries'], )
Add some trove classifiers to the package metadata
## Code Before: import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', ) ## Instruction: Add some trove classifiers to the package metadata ## Code After: import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'License :: DFSG approved', 'Programming Language :: Python', 'Topic :: Internet :: WWW/HTTP :: Indexing/Search', 'Topic :: Software Development :: Libraries'], )
import distutils.core distutils.core.setup( name='sunburnt', version='0.4', description='Python interface to Solr', author='Toby White', author_email='toby@timetric.com', packages=['sunburnt'], requires=['httplib2', 'lxml', 'pytz'], license='WTFPL', + classifiers=[ + 'Development Status :: 3 - Alpha', + 'Intended Audience :: Developers', + 'License :: DFSG approved', + 'Programming Language :: Python', + 'Topic :: Internet :: WWW/HTTP :: Indexing/Search', + 'Topic :: Software Development :: Libraries'], )
e4a3d3c273a1b7e26e9fdcf7f8da060bf127f27e
examples/django_project/django_project/tests.py
examples/django_project/django_project/tests.py
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
Add note for test_api_me django-project test
Add note for test_api_me django-project test
Python
mit
syrusakbary/snapshottest
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): + # Note this tests should fail unless the snapshot-update command line + # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
Add note for test_api_me django-project test
## Code Before: import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main() ## Instruction: Add note for test_api_me django-project test ## Code After: import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): # Note this tests should fail unless the snapshot-update command line # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
import unittest from datetime import datetime from snapshottest.django import TestCase def api_client_get(url): return { 'url': url, } class TestDemo(TestCase): def test_api_me(self): + # Note this tests should fail unless the snapshot-update command line + # option is specified. Run `python manage.py test --snapshot-update`. now = datetime.now().isoformat() my_api_response = api_client_get('/' + now) self.assertMatchSnapshot(my_api_response) if __name__ == '__main__': unittest.main()
69fc2eccaa88189fd0de86d11206fa24d1508819
tools/np_suppressions.py
tools/np_suppressions.py
suppressions = [ [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ], # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. [ ".*/multiarray/calculation\.", "PyArray_Std" ], # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. [ ".*/multiarray/common\.", "PyCapsule_Check" ], ]
suppressions = [ # This one cannot be covered by any Python language test because there is # no code pathway to it. But it is part of the C API, so must not be # excised from the code. [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ], # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. [ r".*/multiarray/calculation\.", "PyArray_Std" ], # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. [ r".*/multiarray/common\.", "PyCapsule_Check" ], ]
Add documentation on one assertion, convert RE's to raw strings.
Add documentation on one assertion, convert RE's to raw strings.
Python
bsd-3-clause
teoliphant/numpy-refactor,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,jasonmccampbell/numpy-refactor-sprint,jasonmccampbell/numpy-refactor-sprint,teoliphant/numpy-refactor,teoliphant/numpy-refactor
suppressions = [ + # This one cannot be covered by any Python language test because there is + # no code pathway to it. But it is part of the C API, so must not be + # excised from the code. - [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ], + [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ], # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. - [ ".*/multiarray/calculation\.", "PyArray_Std" ], + [ r".*/multiarray/calculation\.", "PyArray_Std" ], # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. - [ ".*/multiarray/common\.", "PyCapsule_Check" ], + [ r".*/multiarray/common\.", "PyCapsule_Check" ], ]
Add documentation on one assertion, convert RE's to raw strings.
## Code Before: suppressions = [ [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ], # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. [ ".*/multiarray/calculation\.", "PyArray_Std" ], # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. [ ".*/multiarray/common\.", "PyCapsule_Check" ], ] ## Instruction: Add documentation on one assertion, convert RE's to raw strings. ## Code After: suppressions = [ # This one cannot be covered by any Python language test because there is # no code pathway to it. But it is part of the C API, so must not be # excised from the code. [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ], # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. [ r".*/multiarray/calculation\.", "PyArray_Std" ], # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. [ r".*/multiarray/common\.", "PyCapsule_Check" ], ]
suppressions = [ + # This one cannot be covered by any Python language test because there is + # no code pathway to it. But it is part of the C API, so must not be + # excised from the code. - [ ".*/multiarray/mapping\.", "PyArray_MapIterReset" ], + [ r".*/multiarray/mapping\.", "PyArray_MapIterReset" ], ? + # PyArray_Std trivially forwards to and appears to be superceded by # __New_PyArray_Std, which is exercised by the test framework. - [ ".*/multiarray/calculation\.", "PyArray_Std" ], + [ r".*/multiarray/calculation\.", "PyArray_Std" ], ? + # PyCapsule_Check is declared in a header, and used in # multiarray/ctors.c. So it isn't really untested. - [ ".*/multiarray/common\.", "PyCapsule_Check" ], + [ r".*/multiarray/common\.", "PyCapsule_Check" ], ? + ]
9f3abe5077fce0a2d7323a769fc063fca5b7aca8
tests/test_bawlerd.py
tests/test_bawlerd.py
import os from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list()
import io import os from textwrap import dedent from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list() def test__load_file(self): config = bawlerd.conf._load_file(io.StringIO(dedent("""\ logging: formatters: standard: format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\" handlers: default: level: "INFO" formatter: standard class: logging.StreamHandler loggers: "": handlers: ["default"] level: INFO propagate: True """))) assert 'logging' in config
Add simple test for _load_file
Add simple test for _load_file Signed-off-by: Michal Kuffa <005ee1c97edba97d164343c993afee612ac25a0c@gmail.com>
Python
bsd-3-clause
beezz/pg_bawler,beezz/pg_bawler
+ import io import os + from textwrap import dedent from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list() + def test__load_file(self): + config = bawlerd.conf._load_file(io.StringIO(dedent("""\ + logging: + formatters: + standard: + format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\" + handlers: + default: + level: "INFO" + formatter: standard + class: logging.StreamHandler + loggers: + "": + handlers: ["default"] + level: INFO + propagate: True + """))) + assert 'logging' in config +
Add simple test for _load_file
## Code Before: import os from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list() ## Instruction: Add simple test for _load_file ## Code After: import io import os from textwrap import dedent from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list() def test__load_file(self): config = bawlerd.conf._load_file(io.StringIO(dedent("""\ logging: formatters: standard: format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\" handlers: default: level: "INFO" formatter: standard class: logging.StreamHandler loggers: "": handlers: ["default"] level: INFO propagate: True """))) assert 'logging' in config
+ import io import os + from textwrap import dedent from pg_bawler import bawlerd class TestBawlerdConfig: def test_build_config_location_list(self): assert not bawlerd.conf.build_config_location_list(locations=()) user_conf = os.path.join( os.path.expanduser('~'), bawlerd.conf.DEFAULT_CONFIG_FILENAME) system_conf = os.path.join( '/etc/pg_bawler', bawlerd.conf.DEFAULT_CONFIG_FILENAME) assert user_conf in bawlerd.conf.build_config_location_list() assert system_conf in bawlerd.conf.build_config_location_list() + + def test__load_file(self): + config = bawlerd.conf._load_file(io.StringIO(dedent("""\ + logging: + formatters: + standard: + format: \"%(asctime)s %(levelname)s] %(name)s: %(message)s\" + handlers: + default: + level: "INFO" + formatter: standard + class: logging.StreamHandler + loggers: + "": + handlers: ["default"] + level: INFO + propagate: True + """))) + assert 'logging' in config
1709c602b8a423d1eee6521c5e74987db0fc8b81
fancypages/contrib/oscar_fancypages/mixins.py
fancypages/contrib/oscar_fancypages/mixins.py
from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' context_object_name = 'fancypage' def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) ctx[self.context_object_name] = getattr(self, self.page_attr_name) if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx
from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' context_object_name = 'products' def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) ctx['fancypage'] = getattr(self, self.page_attr_name) if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx
Change context object for product list view in Oscar contrib
Change context object for product list view in Oscar contrib
Python
bsd-3-clause
tangentlabs/django-fancypages,tangentlabs/django-fancypages,tangentlabs/django-fancypages,socradev/django-fancypages,socradev/django-fancypages,socradev/django-fancypages
from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' - context_object_name = 'fancypage' + context_object_name = 'products' def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) - ctx[self.context_object_name] = getattr(self, self.page_attr_name) + ctx['fancypage'] = getattr(self, self.page_attr_name) if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx
Change context object for product list view in Oscar contrib
## Code Before: from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' context_object_name = 'fancypage' def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) ctx[self.context_object_name] = getattr(self, self.page_attr_name) if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx ## Instruction: Change context object for product list view in Oscar contrib ## Code After: from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' context_object_name = 'products' def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) ctx['fancypage'] = getattr(self, self.page_attr_name) if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx
from ... import mixins class OscarFancyPageMixin(mixins.FancyPageMixin): node_attr_name = 'category' slug_url_kwarg = 'category_slug' - context_object_name = 'fancypage' ? ^^^ ^^^^^ + context_object_name = 'products' ? ^^^^^ ^^ def get_context_data(self, **kwargs): ctx = super(OscarFancyPageMixin, self).get_context_data(**kwargs) - ctx[self.context_object_name] = getattr(self, self.page_attr_name) ? ^ ^^^^^^^^^^^^^^^^^^^^^^ + ctx['fancypage'] = getattr(self, self.page_attr_name) ? ^^^^^^^^^ ^ if self.category: ctx['object'] = self.category for container in self.category.page.containers.all(): ctx[container.name] = container return ctx
0dabc858976197459cfe71fe1a4a8a85c181db75
django_localflavor_ie/ie_counties.py
django_localflavor_ie/ie_counties.py
from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( ('antrim', _('Antrim')), ('armagh', _('Armagh')), ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), ('derry', _('Derry')), ('donegal', _('Donegal')), ('down', _('Down')), ('dublin', _('Dublin')), ('fermanagh', _('Fermanagh')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), ('tyrone', _('Tyrone')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), )
from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), ('donegal', _('Donegal')), ('dublin', _('Dublin')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), )
Remove Northern Irish counties. These are part of the UK, not Ireland
Remove Northern Irish counties. These are part of the UK, not Ireland
Python
bsd-3-clause
martinogden/django-localflavor-ie
from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( - ('antrim', _('Antrim')), - ('armagh', _('Armagh')), ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), - ('derry', _('Derry')), ('donegal', _('Donegal')), - ('down', _('Down')), ('dublin', _('Dublin')), - ('fermanagh', _('Fermanagh')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), - ('tyrone', _('Tyrone')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), )
Remove Northern Irish counties. These are part of the UK, not Ireland
## Code Before: from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( ('antrim', _('Antrim')), ('armagh', _('Armagh')), ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), ('derry', _('Derry')), ('donegal', _('Donegal')), ('down', _('Down')), ('dublin', _('Dublin')), ('fermanagh', _('Fermanagh')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), ('tyrone', _('Tyrone')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), ) ## Instruction: Remove Northern Irish counties. These are part of the UK, not Ireland ## Code After: from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), ('donegal', _('Donegal')), ('dublin', _('Dublin')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), )
from django.utils.translation import ugettext_lazy as _ IE_COUNTY_CHOICES = ( - ('antrim', _('Antrim')), - ('armagh', _('Armagh')), ('carlow', _('Carlow')), ('cavan', _('Cavan')), ('clare', _('Clare')), ('cork', _('Cork')), - ('derry', _('Derry')), ('donegal', _('Donegal')), - ('down', _('Down')), ('dublin', _('Dublin')), - ('fermanagh', _('Fermanagh')), ('galway', _('Galway')), ('kerry', _('Kerry')), ('kildare', _('Kildare')), ('kilkenny', _('Kilkenny')), ('laois', _('Laois')), ('leitrim', _('Leitrim')), ('limerick', _('Limerick')), ('longford', _('Longford')), ('louth', _('Louth')), ('mayo', _('Mayo')), ('meath', _('Meath')), ('monaghan', _('Monaghan')), ('offaly', _('Offaly')), ('roscommon', _('Roscommon')), ('sligo', _('Sligo')), ('tipperary', _('Tipperary')), - ('tyrone', _('Tyrone')), ('waterford', _('Waterford')), ('westmeath', _('Westmeath')), ('wexford', _('Wexford')), ('wicklow', _('Wicklow')), )
dab7eaadbc6fc0dd867358b096a846ec39bc0440
pnnl/models/__init__.py
pnnl/models/__init__.py
import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied)
import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied) return q
Add return statement to Model.get_q
Add return statement to Model.get_q
Python
bsd-3-clause
VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications,VOLTTRON/volttron-applications
import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied) + return q
Add return statement to Model.get_q
## Code Before: import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied) ## Instruction: Add return statement to Model.get_q ## Code After: import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied) return q
import importlib import logging from volttron.platform.agent import utils _log = logging.getLogger(__name__) utils.setup_logging() __version__ = "0.1" __all__ = ['Model'] class Model(object): def __init__(self, config, **kwargs): base_module = "volttron.pnnl.models." try: model_type = config["model_type"] except KeyError as e: _log.exception("Missing Model Type key: {}".format(e)) raise e _file, model_type = model_type.split(".") module = importlib.import_module(base_module + _file) model_class = getattr(module, model_type) self.model = model_class(config, self) def get_q(self, _set, sched_index, market_index, occupied): q = self.model.predict(_set, sched_index, market_index, occupied) + return q
e0d631b4aab431c31689ccd7aa6ac92d95e32e80
tests/test_frontend.py
tests/test_frontend.py
import os from tvrenamr.cli import helpers from .base import BaseTest class TestFrontEnd(BaseTest): def setup(self): super(TestFrontEnd, self).setup() self.config = helpers.get_config() def test_passing_current_dir_makes_file_list_a_list(self): assert isinstance(helpers.build_file_list([self.files]), list) def test_setting_recursive_adds_all_files_below_the_folder(self): new_folders = ('herp', 'derp', 'test') os.makedirs(os.path.join(self.files, *new_folders)) def build_folder(folder): new_files = ('foo', 'bar', 'blah') for fn in new_files: with open(os.path.join(self.files, folder, fn), 'w') as f: f.write('') build_folder('herp') build_folder('herp/derp') build_folder('herp/derp/test') file_list = helpers.build_file_list([self.files], recursive=True) for root, dirs, files in os.walk(self.files): for fn in files: assert os.path.join(root, fn) in file_list def test_ignoring_files(self): ignore = self.random_files(self.files) file_list = helpers.build_file_list([self.files], ignore_filelist=ignore) assert all(fn not in file_list for fn in ignore)
import collections import os import sys from tvrenamr.cli import helpers from .utils import random_files def test_passing_current_dir_makes_file_list_a_list(files): file_list = helpers.build_file_list([files]) assert isinstance(file_list, collections.Iterable) PY3 = sys.version_info[0] == 3 string_type = str if PY3 else basestring text_type = str if PY3 else unicode assert not isinstance(file_list, (string_type, text_type)) def test_setting_recursive_adds_all_files_below_the_folder(files): new_folders = ('herp', 'derp', 'test') os.makedirs(os.path.join(files, *new_folders)) def build_folder(folder): new_files = ('foo', 'bar', 'blah') for fn in new_files: with open(os.path.join(files, folder, fn), 'w') as f: f.write('') build_folder('herp') build_folder('herp/derp') build_folder('herp/derp/test') file_list = helpers.build_file_list([files], recursive=True) for root, dirs, files in os.walk(files): for fn in files: assert (root, fn) in file_list def test_ignoring_files(files): ignore = random_files(files) file_list = helpers.build_file_list([files], ignore_filelist=ignore) assert all(fn not in file_list for fn in ignore)
Move to function only tests & fix test for generator based build_file_list
Move to function only tests & fix test for generator based build_file_list build_file_list is a generator now so we need to make sure it returns an iterable but not a string.
Python
mit
wintersandroid/tvrenamr,ghickman/tvrenamr
+ import collections import os + import sys from tvrenamr.cli import helpers - from .base import BaseTest + from .utils import random_files + def test_passing_current_dir_makes_file_list_a_list(files): + file_list = helpers.build_file_list([files]) - class TestFrontEnd(BaseTest): - def setup(self): - super(TestFrontEnd, self).setup() - self.config = helpers.get_config() + assert isinstance(file_list, collections.Iterable) - def test_passing_current_dir_makes_file_list_a_list(self): - assert isinstance(helpers.build_file_list([self.files]), list) - def test_setting_recursive_adds_all_files_below_the_folder(self): - new_folders = ('herp', 'derp', 'test') - os.makedirs(os.path.join(self.files, *new_folders)) + PY3 = sys.version_info[0] == 3 + string_type = str if PY3 else basestring + text_type = str if PY3 else unicode + assert not isinstance(file_list, (string_type, text_type)) - def build_folder(folder): - new_files = ('foo', 'bar', 'blah') - for fn in new_files: - with open(os.path.join(self.files, folder, fn), 'w') as f: - f.write('') - build_folder('herp') - build_folder('herp/derp') - build_folder('herp/derp/test') - file_list = helpers.build_file_list([self.files], recursive=True) - for root, dirs, files in os.walk(self.files): - for fn in files: - assert os.path.join(root, fn) in file_list + def test_setting_recursive_adds_all_files_below_the_folder(files): + new_folders = ('herp', 'derp', 'test') + os.makedirs(os.path.join(files, *new_folders)) - def test_ignoring_files(self): - ignore = self.random_files(self.files) - file_list = helpers.build_file_list([self.files], ignore_filelist=ignore) - assert all(fn not in file_list for fn in ignore) + def build_folder(folder): + new_files = ('foo', 'bar', 'blah') + for fn in new_files: + with open(os.path.join(files, folder, fn), 'w') as f: + f.write('') + build_folder('herp') + build_folder('herp/derp') + build_folder('herp/derp/test') + file_list = helpers.build_file_list([files], recursive=True) + for root, dirs, files in os.walk(files): + for fn in files: + assert (root, fn) in file_list + + + def test_ignoring_files(files): + ignore = random_files(files) + file_list = helpers.build_file_list([files], ignore_filelist=ignore) + assert all(fn not in file_list for fn in ignore) +
Move to function only tests & fix test for generator based build_file_list
## Code Before: import os from tvrenamr.cli import helpers from .base import BaseTest class TestFrontEnd(BaseTest): def setup(self): super(TestFrontEnd, self).setup() self.config = helpers.get_config() def test_passing_current_dir_makes_file_list_a_list(self): assert isinstance(helpers.build_file_list([self.files]), list) def test_setting_recursive_adds_all_files_below_the_folder(self): new_folders = ('herp', 'derp', 'test') os.makedirs(os.path.join(self.files, *new_folders)) def build_folder(folder): new_files = ('foo', 'bar', 'blah') for fn in new_files: with open(os.path.join(self.files, folder, fn), 'w') as f: f.write('') build_folder('herp') build_folder('herp/derp') build_folder('herp/derp/test') file_list = helpers.build_file_list([self.files], recursive=True) for root, dirs, files in os.walk(self.files): for fn in files: assert os.path.join(root, fn) in file_list def test_ignoring_files(self): ignore = self.random_files(self.files) file_list = helpers.build_file_list([self.files], ignore_filelist=ignore) assert all(fn not in file_list for fn in ignore) ## Instruction: Move to function only tests & fix test for generator based build_file_list ## Code After: import collections import os import sys from tvrenamr.cli import helpers from .utils import random_files def test_passing_current_dir_makes_file_list_a_list(files): file_list = helpers.build_file_list([files]) assert isinstance(file_list, collections.Iterable) PY3 = sys.version_info[0] == 3 string_type = str if PY3 else basestring text_type = str if PY3 else unicode assert not isinstance(file_list, (string_type, text_type)) def test_setting_recursive_adds_all_files_below_the_folder(files): new_folders = ('herp', 'derp', 'test') os.makedirs(os.path.join(files, *new_folders)) def build_folder(folder): new_files = ('foo', 'bar', 'blah') for fn in new_files: with open(os.path.join(files, folder, fn), 'w') as f: f.write('') build_folder('herp') build_folder('herp/derp') build_folder('herp/derp/test') file_list = helpers.build_file_list([files], recursive=True) for root, dirs, files in os.walk(files): for fn in files: assert (root, fn) in file_list def test_ignoring_files(files): ignore = random_files(files) file_list = helpers.build_file_list([files], ignore_filelist=ignore) assert all(fn not in file_list for fn in ignore)
+ import collections import os + import sys from tvrenamr.cli import helpers - from .base import BaseTest + from .utils import random_files + def test_passing_current_dir_makes_file_list_a_list(files): + file_list = helpers.build_file_list([files]) - class TestFrontEnd(BaseTest): - def setup(self): - super(TestFrontEnd, self).setup() - self.config = helpers.get_config() + assert isinstance(file_list, collections.Iterable) - def test_passing_current_dir_makes_file_list_a_list(self): - assert isinstance(helpers.build_file_list([self.files]), list) - def test_setting_recursive_adds_all_files_below_the_folder(self): - new_folders = ('herp', 'derp', 'test') - os.makedirs(os.path.join(self.files, *new_folders)) + PY3 = sys.version_info[0] == 3 + string_type = str if PY3 else basestring + text_type = str if PY3 else unicode + assert not isinstance(file_list, (string_type, text_type)) - def build_folder(folder): - new_files = ('foo', 'bar', 'blah') - for fn in new_files: - with open(os.path.join(self.files, folder, fn), 'w') as f: - f.write('') - build_folder('herp') - build_folder('herp/derp') - build_folder('herp/derp/test') - file_list = helpers.build_file_list([self.files], recursive=True) - for root, dirs, files in os.walk(self.files): - for fn in files: - assert os.path.join(root, fn) in file_list + def test_setting_recursive_adds_all_files_below_the_folder(files): + new_folders = ('herp', 'derp', 'test') + os.makedirs(os.path.join(files, *new_folders)) + + def build_folder(folder): + new_files = ('foo', 'bar', 'blah') + for fn in new_files: + with open(os.path.join(files, folder, fn), 'w') as f: + f.write('') + build_folder('herp') + build_folder('herp/derp') + build_folder('herp/derp/test') + file_list = helpers.build_file_list([files], recursive=True) + for root, dirs, files in os.walk(files): + for fn in files: + assert (root, fn) in file_list + + - def test_ignoring_files(self): ? ---- --- + def test_ignoring_files(files): ? ++++ - ignore = self.random_files(self.files) ? ---- ----- ----- + ignore = random_files(files) - file_list = helpers.build_file_list([self.files], ignore_filelist=ignore) ? ---- ----- + file_list = helpers.build_file_list([files], ignore_filelist=ignore) - assert all(fn not in file_list for fn in ignore) ? ---- + assert all(fn not in file_list for fn in ignore)
e2d51e23f530202b82ba13ae11c686deb1388435
prototype/BioID.py
prototype/BioID.py
import re import json import mmap class BioID: defs = None def __init__(self, defpath): with open(defpath, "r") as deffile: conts = deffile.read() self.defs = json.loads(conts)["formats"] @classmethod def identify(cls, files): recog = {} for file in files: with open(file, "r") as infile: buff = infile.read() mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) if len(buff) == 0: recog[file] = "empty" # Empty files have no format :) continue for fdef in cls.defs: matched = True if "regexen" in fdef: for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): matched = False break if "bytes" in fdef: for bytes in fdef["bytes"]: if mem_map.find(bytes.decode("string_escape")) == -1: matched = False break if matched: recog[file] = fdef["name"] break mem_map.close() if file not in recog: recog[file] = "unrecognized" return recog
import re import json import mmap class BioID: defs = None def __init__(self, defpath): with open(defpath, "r") as deffile: conts = deffile.read() self.defs = json.loads(conts)["formats"] @classmethod def identify(cls, files): recog = {} for file in files: with open(file, "r") as infile: buff = infile.read() mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) if len(buff) == 0: recog[file] = "empty" # Empty files have no format :) continue for fdef in cls.defs: matched = True if "regexen" in fdef: for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): matched = False break if "bytes" in fdef: for bytes in fdef["bytes"]: if mem_map.find(bytes.decode("string_escape")) == -1: matched = False break if matched: recog[file] = fdef["name"] break mem_map.close() if file not in recog: recog[file] = "unrecognized" return recog
Indent return in identify class.
Indent return in identify class.
Python
mit
LeeBergstrand/BioMagick,LeeBergstrand/BioMagick
import re import json import mmap class BioID: - defs = None + defs = None - def __init__(self, defpath): + def __init__(self, defpath): - with open(defpath, "r") as deffile: + with open(defpath, "r") as deffile: - conts = deffile.read() + conts = deffile.read() - self.defs = json.loads(conts)["formats"] + self.defs = json.loads(conts)["formats"] - @classmethod + @classmethod - def identify(cls, files): + def identify(cls, files): - recog = {} + recog = {} - for file in files: + for file in files: - with open(file, "r") as infile: + with open(file, "r") as infile: - buff = infile.read() + buff = infile.read() - mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) + mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) - if len(buff) == 0: + if len(buff) == 0: - recog[file] = "empty" # Empty files have no format :) + recog[file] = "empty" # Empty files have no format :) - continue + continue - for fdef in cls.defs: - matched = True - if "regexen" in fdef: - for regex in fdef["regexen"]: + for fdef in cls.defs: + matched = True + if "regexen" in fdef: + for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): - matched = False - break - if "bytes" in fdef: - for bytes in fdef["bytes"]: + matched = False + break + if "bytes" in fdef: + for bytes in fdef["bytes"]: - if mem_map.find(bytes.decode("string_escape")) == -1: + if mem_map.find(bytes.decode("string_escape")) == -1: - matched = False - break - if matched: - recog[file] = fdef["name"] - break + matched = False + break + if matched: + recog[file] = fdef["name"] + break - mem_map.close() + mem_map.close() - if file not in recog: + if file not in recog: - recog[file] = "unrecognized" + recog[file] = "unrecognized" - return recog + return recog
Indent return in identify class.
## Code Before: import re import json import mmap class BioID: defs = None def __init__(self, defpath): with open(defpath, "r") as deffile: conts = deffile.read() self.defs = json.loads(conts)["formats"] @classmethod def identify(cls, files): recog = {} for file in files: with open(file, "r") as infile: buff = infile.read() mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) if len(buff) == 0: recog[file] = "empty" # Empty files have no format :) continue for fdef in cls.defs: matched = True if "regexen" in fdef: for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): matched = False break if "bytes" in fdef: for bytes in fdef["bytes"]: if mem_map.find(bytes.decode("string_escape")) == -1: matched = False break if matched: recog[file] = fdef["name"] break mem_map.close() if file not in recog: recog[file] = "unrecognized" return recog ## Instruction: Indent return in identify class. ## Code After: import re import json import mmap class BioID: defs = None def __init__(self, defpath): with open(defpath, "r") as deffile: conts = deffile.read() self.defs = json.loads(conts)["formats"] @classmethod def identify(cls, files): recog = {} for file in files: with open(file, "r") as infile: buff = infile.read() mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) if len(buff) == 0: recog[file] = "empty" # Empty files have no format :) continue for fdef in cls.defs: matched = True if "regexen" in fdef: for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): matched = False break if "bytes" in fdef: for bytes in fdef["bytes"]: if mem_map.find(bytes.decode("string_escape")) == -1: matched = False break if matched: recog[file] = fdef["name"] break mem_map.close() if file not in recog: recog[file] = "unrecognized" return recog
import re import json import mmap class BioID: - defs = None ? ^^^^ + defs = None ? ^ - def __init__(self, defpath): ? ^^^^ + def __init__(self, defpath): ? ^ - with open(defpath, "r") as deffile: ? ^^^^^^^^ + with open(defpath, "r") as deffile: ? ^^ - conts = deffile.read() + conts = deffile.read() - self.defs = json.loads(conts)["formats"] ? ^^^^^^^^ + self.defs = json.loads(conts)["formats"] ? ^^ - @classmethod ? ^^^^ + @classmethod ? ^ - def identify(cls, files): ? ^^^^ + def identify(cls, files): ? ^ - recog = {} + recog = {} - for file in files: ? ^^^^^^^^ + for file in files: ? ^^ - with open(file, "r") as infile: ? ^^^^^^^^^^^^ + with open(file, "r") as infile: ? ^^^ - buff = infile.read() + buff = infile.read() - mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) ? ^^^^^^^^^^^^^^^^ + mem_map = mmap.mmap(infile.fileno(), 0, mmap.MAP_PRIVATE, mmap.PROT_READ) ? ^^^^ - if len(buff) == 0: + if len(buff) == 0: - recog[file] = "empty" # Empty files have no format :) ? ^^^^^^^^^^^^^^^^ + recog[file] = "empty" # Empty files have no format :) ? ^^^^ - continue + continue - for fdef in cls.defs: - matched = True - if "regexen" in fdef: - for regex in fdef["regexen"]: + for fdef in cls.defs: + matched = True + if "regexen" in fdef: + for regex in fdef["regexen"]: if not re.findall(regex.replace("\\n", "\n"), buff, re.IGNORECASE): - matched = False - break - if "bytes" in fdef: - for bytes in fdef["bytes"]: + matched = False + break + if "bytes" in fdef: + for bytes in fdef["bytes"]: - if mem_map.find(bytes.decode("string_escape")) == -1: ? ^^^^^^^^^^^^^^^^^^^^^^^^ + if mem_map.find(bytes.decode("string_escape")) == -1: ? ^^^^^ - matched = False - break - if matched: - recog[file] = fdef["name"] - break + matched = False + break + if matched: + recog[file] = fdef["name"] + break - mem_map.close() + mem_map.close() - if file not in recog: ? ^^^^^^^^^^^^ + if file not in recog: ? ^^ - recog[file] = "unrecognized" + recog[file] = "unrecognized" - return recog + return recog
c814fe264c93dfa09276474960aa83cdb26e7754
polyaxon/api/searches/serializers.py
polyaxon/api/searches/serializers.py
from rest_framework import serializers from db.models.searches import Search class SearchSerializer(serializers.ModelSerializer): class Meta: model = Search fields = ['id', 'name', 'query', 'meta']
from rest_framework import serializers from rest_framework.exceptions import ValidationError from api.utils.serializers.names import NamesMixin from db.models.searches import Search class SearchSerializer(serializers.ModelSerializer, NamesMixin): class Meta: model = Search fields = ['id', 'name', 'query', 'meta'] def create(self, validated_data): validated_data = self.validated_name(validated_data, project=validated_data['project'], query=Search.all) try: return super().create(validated_data) except Exception as e: raise ValidationError(e)
Add graceful handling for creating search with similar names
Add graceful handling for creating search with similar names
Python
apache-2.0
polyaxon/polyaxon,polyaxon/polyaxon,polyaxon/polyaxon
from rest_framework import serializers + from rest_framework.exceptions import ValidationError + from api.utils.serializers.names import NamesMixin from db.models.searches import Search - class SearchSerializer(serializers.ModelSerializer): + class SearchSerializer(serializers.ModelSerializer, NamesMixin): - class Meta: model = Search fields = ['id', 'name', 'query', 'meta'] + def create(self, validated_data): + validated_data = self.validated_name(validated_data, + project=validated_data['project'], + query=Search.all) + try: + return super().create(validated_data) + except Exception as e: + raise ValidationError(e) +
Add graceful handling for creating search with similar names
## Code Before: from rest_framework import serializers from db.models.searches import Search class SearchSerializer(serializers.ModelSerializer): class Meta: model = Search fields = ['id', 'name', 'query', 'meta'] ## Instruction: Add graceful handling for creating search with similar names ## Code After: from rest_framework import serializers from rest_framework.exceptions import ValidationError from api.utils.serializers.names import NamesMixin from db.models.searches import Search class SearchSerializer(serializers.ModelSerializer, NamesMixin): class Meta: model = Search fields = ['id', 'name', 'query', 'meta'] def create(self, validated_data): validated_data = self.validated_name(validated_data, project=validated_data['project'], query=Search.all) try: return super().create(validated_data) except Exception as e: raise ValidationError(e)
from rest_framework import serializers + from rest_framework.exceptions import ValidationError + from api.utils.serializers.names import NamesMixin from db.models.searches import Search - class SearchSerializer(serializers.ModelSerializer): + class SearchSerializer(serializers.ModelSerializer, NamesMixin): ? ++++++++++++ - class Meta: model = Search fields = ['id', 'name', 'query', 'meta'] + + def create(self, validated_data): + validated_data = self.validated_name(validated_data, + project=validated_data['project'], + query=Search.all) + try: + return super().create(validated_data) + except Exception as e: + raise ValidationError(e)
0a2fa84285a586282d79146f85d9efba12a528dd
Parallel/Testing/Cxx/TestSockets.py
Parallel/Testing/Cxx/TestSockets.py
import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15)
import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) sys.exit(os.WEXITSTATUS(retVal))
Return code from script must reflect that of the test.
BUG: Return code from script must reflect that of the test.
Python
bsd-3-clause
mspark93/VTK,jeffbaumes/jeffbaumes-vtk,demarle/VTK,sumedhasingla/VTK,sankhesh/VTK,mspark93/VTK,keithroe/vtkoptix,daviddoria/PointGraphsPhase1,SimVascular/VTK,collects/VTK,SimVascular/VTK,sumedhasingla/VTK,mspark93/VTK,biddisco/VTK,sgh/vtk,jmerkow/VTK,aashish24/VTK-old,demarle/VTK,demarle/VTK,aashish24/VTK-old,johnkit/vtk-dev,johnkit/vtk-dev,johnkit/vtk-dev,msmolens/VTK,hendradarwin/VTK,SimVascular/VTK,johnkit/vtk-dev,hendradarwin/VTK,cjh1/VTK,demarle/VTK,ashray/VTK-EVM,daviddoria/PointGraphsPhase1,sankhesh/VTK,keithroe/vtkoptix,Wuteyan/VTK,mspark93/VTK,spthaolt/VTK,candy7393/VTK,aashish24/VTK-old,jmerkow/VTK,sgh/vtk,gram526/VTK,Wuteyan/VTK,daviddoria/PointGraphsPhase1,sgh/vtk,spthaolt/VTK,keithroe/vtkoptix,sumedhasingla/VTK,hendradarwin/VTK,berendkleinhaneveld/VTK,ashray/VTK-EVM,berendkleinhaneveld/VTK,jeffbaumes/jeffbaumes-vtk,jmerkow/VTK,sankhesh/VTK,SimVascular/VTK,ashray/VTK-EVM,keithroe/vtkoptix,ashray/VTK-EVM,arnaudgelas/VTK,biddisco/VTK,candy7393/VTK,collects/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,jeffbaumes/jeffbaumes-vtk,naucoin/VTKSlicerWidgets,aashish24/VTK-old,Wuteyan/VTK,msmolens/VTK,jmerkow/VTK,cjh1/VTK,jeffbaumes/jeffbaumes-vtk,naucoin/VTKSlicerWidgets,jmerkow/VTK,johnkit/vtk-dev,msmolens/VTK,jmerkow/VTK,spthaolt/VTK,cjh1/VTK,spthaolt/VTK,msmolens/VTK,biddisco/VTK,cjh1/VTK,collects/VTK,arnaudgelas/VTK,Wuteyan/VTK,aashish24/VTK-old,spthaolt/VTK,hendradarwin/VTK,mspark93/VTK,naucoin/VTKSlicerWidgets,naucoin/VTKSlicerWidgets,jeffbaumes/jeffbaumes-vtk,SimVascular/VTK,Wuteyan/VTK,sankhesh/VTK,sankhesh/VTK,hendradarwin/VTK,arnaudgelas/VTK,daviddoria/PointGraphsPhase1,collects/VTK,spthaolt/VTK,Wuteyan/VTK,johnkit/vtk-dev,Wuteyan/VTK,gram526/VTK,candy7393/VTK,candy7393/VTK,candy7393/VTK,keithroe/vtkoptix,keithroe/vtkoptix,sumedhasingla/VTK,sankhesh/VTK,berendkleinhaneveld/VTK,candy7393/VTK,msmolens/VTK,demarle/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,msmolens/VTK,ashray/VTK-EVM,daviddoria/PointGraphsPhase1,msmolens/VTK,keithroe/vtkoptix,gram526/VTK,sgh/vtk,sgh/vtk,keithroe/vtkoptix,berendkleinhaneveld/VTK,demarle/VTK,sankhesh/VTK,candy7393/VTK,daviddoria/PointGraphsPhase1,collects/VTK,cjh1/VTK,jeffbaumes/jeffbaumes-vtk,collects/VTK,johnkit/vtk-dev,gram526/VTK,sumedhasingla/VTK,biddisco/VTK,ashray/VTK-EVM,mspark93/VTK,mspark93/VTK,sgh/vtk,demarle/VTK,arnaudgelas/VTK,hendradarwin/VTK,jmerkow/VTK,gram526/VTK,candy7393/VTK,SimVascular/VTK,jmerkow/VTK,aashish24/VTK-old,gram526/VTK,biddisco/VTK,mspark93/VTK,sankhesh/VTK,hendradarwin/VTK,ashray/VTK-EVM,biddisco/VTK,SimVascular/VTK,msmolens/VTK,naucoin/VTKSlicerWidgets,biddisco/VTK,gram526/VTK,cjh1/VTK,ashray/VTK-EVM,SimVascular/VTK,naucoin/VTKSlicerWidgets,arnaudgelas/VTK,arnaudgelas/VTK,spthaolt/VTK,gram526/VTK,sumedhasingla/VTK,berendkleinhaneveld/VTK,demarle/VTK
import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client - os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], + retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], - sys.argv[4] )) + sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) + sys.exit(os.WEXITSTATUS(retVal)) -
Return code from script must reflect that of the test.
## Code Before: import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) ## Instruction: Return code from script must reflect that of the test. ## Code After: import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], sys.argv[4] )) # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) sys.exit(os.WEXITSTATUS(retVal))
import os, sys, time # Fork, run server in child, client in parent pid = os.fork() if pid == 0: # exec the parent os.execv(sys.argv[1], ('-D', sys.argv[3])) else: # wait a little to make sure that the server is ready time.sleep(10) # run the client - os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], + retVal = os.system('%s -D %s -V %s' % ( sys.argv[2], sys.argv[3], ? +++++++++ - sys.argv[4] )) + sys.argv[4] )) ? +++++++++ # in case the client fails, we need to kill the server # or it will stay around time.sleep(20) os.kill(pid, 15) - + sys.exit(os.WEXITSTATUS(retVal))
2f860583a99b88324b19b1118b4aea29a28ae90d
polling_stations/apps/data_collection/management/commands/import_portsmouth.py
polling_stations/apps/data_collection/management/commands/import_portsmouth.py
from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" elections = ["local.2018-05-03"] csv_delimiter = "\t" def address_record_to_dict(self, record): if record.addressline6 == "PO1 5BZ": return None return super().address_record_to_dict(record) def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) if rec["internal_council_id"] == "3270": rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec
from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" addresses_name = ( "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" ) stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" elections = ["local.2019-05-02"] csv_delimiter = "\t" def address_record_to_dict(self, record): rec = super().address_record_to_dict(record) if record.addressline6 == "PO4 099": rec["postcode"] = "PO4 0PL" if record.property_urn.strip().lstrip("0") in [ "1775122942", "1775122943", "1775122944", ]: rec["postcode"] = "PO5 2BZ" return rec def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) if rec["internal_council_id"] == "3596": rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec
Add import script for Portsmouth
Add import script for Portsmouth Closes #1502
Python
bsd-3-clause
DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations,DemocracyClub/UK-Polling-Stations
from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" + addresses_name = ( - addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" + "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" + ) - stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" + stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" - elections = ["local.2018-05-03"] + elections = ["local.2019-05-02"] csv_delimiter = "\t" def address_record_to_dict(self, record): + rec = super().address_record_to_dict(record) - if record.addressline6 == "PO1 5BZ": + if record.addressline6 == "PO4 099": - return None + rec["postcode"] = "PO4 0PL" - return super().address_record_to_dict(record) + if record.property_urn.strip().lstrip("0") in [ + "1775122942", + "1775122943", + "1775122944", + ]: + rec["postcode"] = "PO5 2BZ" + + return rec def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) - if rec["internal_council_id"] == "3270": + if rec["internal_council_id"] == "3596": rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec
Add import script for Portsmouth
## Code Before: from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" elections = ["local.2018-05-03"] csv_delimiter = "\t" def address_record_to_dict(self, record): if record.addressline6 == "PO1 5BZ": return None return super().address_record_to_dict(record) def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) if rec["internal_council_id"] == "3270": rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec ## Instruction: Add import script for Portsmouth ## Code After: from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" addresses_name = ( "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" ) stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" elections = ["local.2019-05-02"] csv_delimiter = "\t" def address_record_to_dict(self, record): rec = super().address_record_to_dict(record) if record.addressline6 == "PO4 099": rec["postcode"] = "PO4 0PL" if record.property_urn.strip().lstrip("0") in [ "1775122942", "1775122943", "1775122944", ]: rec["postcode"] = "PO5 2BZ" return rec def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) if rec["internal_council_id"] == "3596": rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec
from django.contrib.gis.geos import Point from data_collection.management.commands import BaseXpressDemocracyClubCsvImporter class Command(BaseXpressDemocracyClubCsvImporter): council_id = "E06000044" + addresses_name = ( - addresses_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" ? -------------- ^ ^ ^ ^ ^ + "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" ? ^^ ^ ^ ^ ^^^^^^^^^^^ + ) - stations_name = "local.2018-05-03/Version 1/Democracy_Club__03May2018.tsv" ? ^ ^ ^ ^ + stations_name = "local.2019-05-02/Version 1/Democracy_Club__02May2019Portsmouth.tsv" ? ^ ^ ^ ^^^^^^^^^^^ - elections = ["local.2018-05-03"] ? ^ ^ + elections = ["local.2019-05-02"] ? ^ ^ csv_delimiter = "\t" def address_record_to_dict(self, record): + rec = super().address_record_to_dict(record) - if record.addressline6 == "PO1 5BZ": ? ^ ^^^ + if record.addressline6 == "PO4 099": ? ^ ^^^ - return None + rec["postcode"] = "PO4 0PL" - return super().address_record_to_dict(record) + if record.property_urn.strip().lstrip("0") in [ + "1775122942", + "1775122943", + "1775122944", + ]: + rec["postcode"] = "PO5 2BZ" + + return rec def station_record_to_dict(self, record): rec = super().station_record_to_dict(record) - if rec["internal_council_id"] == "3270": ? ^^^ + if rec["internal_council_id"] == "3596": ? ^^^ rec["location"] = Point(-1.059545, 50.7866578, srid=4326) return rec
0d135a746dd79ad1b703570e2bb3b27a694c67b0
simuvex/procedures/stubs/NoReturnUnconstrained.py
simuvex/procedures/stubs/NoReturnUnconstrained.py
import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'} class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin NO_RET = True def run(self): #pylint:disable=unused-argument return
import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp', '__longjmp_chk', '__siglongjmp_chk'} NO_RET = True def run(self, **kwargs): #pylint:disable=unused-argument return
Move use_cases into the class
Move use_cases into the class
Python
bsd-2-clause
iamahuman/angr,iamahuman/angr,f-prettyland/angr,chubbymaggie/simuvex,chubbymaggie/angr,schieb/angr,angr/angr,tyb0807/angr,f-prettyland/angr,tyb0807/angr,axt/angr,tyb0807/angr,chubbymaggie/angr,angr/angr,angr/angr,iamahuman/angr,axt/angr,f-prettyland/angr,chubbymaggie/angr,angr/simuvex,chubbymaggie/simuvex,zhuyue1314/simuvex,axt/angr,schieb/angr,chubbymaggie/simuvex,schieb/angr
import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### - use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'} - class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin + use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp', + '__longjmp_chk', '__siglongjmp_chk'} NO_RET = True - def run(self): #pylint:disable=unused-argument + def run(self, **kwargs): #pylint:disable=unused-argument return
Move use_cases into the class
## Code Before: import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'} class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin NO_RET = True def run(self): #pylint:disable=unused-argument return ## Instruction: Move use_cases into the class ## Code After: import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp', '__longjmp_chk', '__siglongjmp_chk'} NO_RET = True def run(self, **kwargs): #pylint:disable=unused-argument return
import simuvex ###################################### # NoReturnUnconstrained # Use in places you would put ReturnUnconstrained as a default action # But the function shouldn't actually return ###################################### - use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp'} - class NoReturnUnconstrained(simuvex.SimProcedure): #pylint:disable=redefined-builtin + use_cases = {'exit_group', 'exit', 'abort', 'longjmp', 'pthread_exit', 'siglongjmp', + '__longjmp_chk', '__siglongjmp_chk'} NO_RET = True - def run(self): #pylint:disable=unused-argument + def run(self, **kwargs): #pylint:disable=unused-argument ? ++++++++++ return
27a33628310cbd68632f0e8b514de731a033f8e6
IPython/utils/tests/test_shimmodule.py
IPython/utils/tests/test_shimmodule.py
import sys import warnings from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) with warnings.catch_warnings(record=True) as w: warnings.simplefilter("always") import IPython.config assert len(w) == 1 assert issubclass(w[-1].category, ShimWarning)
import pytest import sys from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) with pytest.warns(ShimWarning): import IPython.config
Make test_shim_warning not fail on unrelated warnings
Make test_shim_warning not fail on unrelated warnings
Python
bsd-3-clause
ipython/ipython,ipython/ipython
+ import pytest import sys - import warnings from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) + with pytest.warns(ShimWarning): - with warnings.catch_warnings(record=True) as w: - warnings.simplefilter("always") import IPython.config - assert len(w) == 1 - assert issubclass(w[-1].category, ShimWarning)
Make test_shim_warning not fail on unrelated warnings
## Code Before: import sys import warnings from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) with warnings.catch_warnings(record=True) as w: warnings.simplefilter("always") import IPython.config assert len(w) == 1 assert issubclass(w[-1].category, ShimWarning) ## Instruction: Make test_shim_warning not fail on unrelated warnings ## Code After: import pytest import sys from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) with pytest.warns(ShimWarning): import IPython.config
+ import pytest import sys - import warnings from IPython.utils.shimmodule import ShimWarning def test_shim_warning(): sys.modules.pop('IPython.config', None) + with pytest.warns(ShimWarning): - with warnings.catch_warnings(record=True) as w: - warnings.simplefilter("always") import IPython.config - assert len(w) == 1 - assert issubclass(w[-1].category, ShimWarning)
74816d4af07808009b89163060f97014b1a20ceb
tests/test_arguments.py
tests/test_arguments.py
import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', '__cmp__', '__hash__'] @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv'
import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', '__cmp__', '__hash__', '__nonzero__'] @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv'
Enforce that arguments must implement non-zero methods.
Enforce that arguments must implement non-zero methods.
Python
apache-2.0
disqus/gutter,disqus/gutter,kalail/gutter,kalail/gutter,kalail/gutter
import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', - '__cmp__', '__hash__'] + '__cmp__', '__hash__', '__nonzero__'] @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv'
Enforce that arguments must implement non-zero methods.
## Code Before: import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', '__cmp__', '__hash__'] @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv' ## Instruction: Enforce that arguments must implement non-zero methods. ## Code After: import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', '__cmp__', '__hash__', '__nonzero__'] @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv'
import unittest from mock import MagicMock, Mock from nose.tools import * from gargoyle.inputs.arguments import * class BaseArgument(object): def setUp(self): self.argument = self.klass(self.valid_comparison_value) @property def interface_functions(self): return ['__lt__', '__le__', '__eq__', '__ne__', '__gt__', '__ge__', - '__cmp__', '__hash__'] + '__cmp__', '__hash__', '__nonzero__'] ? +++++++++++++++ @property def interface_methods(self): return [getattr(self.argument, f) for f in self.interface_functions] def test_implements_comparison_methods(self): map(ok_, self.interface_methods) class DelegateToValue(object): def test_delegates_all_interface_function_to_the_value_passed_in(self): value_passed_in = MagicMock() value_passed_in.__cmp__ = Mock() argument = self.klass(value_passed_in) for function in self.interface_functions: values_function = getattr(value_passed_in, function) arguments_function = getattr(argument, function) arguments_function(self.valid_comparison_value) values_function.assert_called_once_with(self.valid_comparison_value) class ValueTest(BaseArgument, DelegateToValue, unittest.TestCase): klass = Value @property def valid_comparison_value(self): return 'marv'
61a4743b62914559fea18a945f7a780e1394da2f
test/test_export_flow.py
test/test_export_flow.py
import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', headers=None, content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) def test_request_simple(): flow = tutils.tflow(req=req_get) assert flow_export.curl_command(flow) flow = tutils.tflow(req=req_post) assert flow_export.curl_command(flow)
import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) req_patch = netlib.tutils.treq( method='PATCH', path=b"/path?query=param", ) def test_curl_command(): flow = tutils.tflow(req=req_get) result = """curl -H 'header:qvalue' 'http://address/path'""" assert flow_export.curl_command(flow) == result flow = tutils.tflow(req=req_post) result = """curl -X POST 'http://address/path' --data-binary 'content'""" assert flow_export.curl_command(flow) == result flow = tutils.tflow(req=req_patch) result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'""" assert flow_export.curl_command(flow) == result
Test exact return value of flow_export.curl_command
Test exact return value of flow_export.curl_command
Python
mit
jvillacorta/mitmproxy,tdickers/mitmproxy,ddworken/mitmproxy,StevenVanAcker/mitmproxy,cortesi/mitmproxy,vhaupert/mitmproxy,tdickers/mitmproxy,mosajjal/mitmproxy,mosajjal/mitmproxy,fimad/mitmproxy,fimad/mitmproxy,ujjwal96/mitmproxy,vhaupert/mitmproxy,dwfreed/mitmproxy,ParthGanatra/mitmproxy,xaxa89/mitmproxy,mhils/mitmproxy,mhils/mitmproxy,tdickers/mitmproxy,StevenVanAcker/mitmproxy,mitmproxy/mitmproxy,ikoz/mitmproxy,dufferzafar/mitmproxy,mitmproxy/mitmproxy,tdickers/mitmproxy,vhaupert/mitmproxy,StevenVanAcker/mitmproxy,Kriechi/mitmproxy,zlorb/mitmproxy,gzzhanghao/mitmproxy,ikoz/mitmproxy,MatthewShao/mitmproxy,mhils/mitmproxy,mosajjal/mitmproxy,gzzhanghao/mitmproxy,mitmproxy/mitmproxy,xaxa89/mitmproxy,cortesi/mitmproxy,ujjwal96/mitmproxy,jvillacorta/mitmproxy,ddworken/mitmproxy,MatthewShao/mitmproxy,cortesi/mitmproxy,laurmurclar/mitmproxy,zlorb/mitmproxy,dwfreed/mitmproxy,dwfreed/mitmproxy,laurmurclar/mitmproxy,dwfreed/mitmproxy,mitmproxy/mitmproxy,mosajjal/mitmproxy,dufferzafar/mitmproxy,Kriechi/mitmproxy,dufferzafar/mitmproxy,ikoz/mitmproxy,jvillacorta/mitmproxy,ikoz/mitmproxy,mhils/mitmproxy,MatthewShao/mitmproxy,gzzhanghao/mitmproxy,ujjwal96/mitmproxy,ujjwal96/mitmproxy,gzzhanghao/mitmproxy,cortesi/mitmproxy,laurmurclar/mitmproxy,ParthGanatra/mitmproxy,mitmproxy/mitmproxy,zlorb/mitmproxy,fimad/mitmproxy,zlorb/mitmproxy,xaxa89/mitmproxy,mhils/mitmproxy,StevenVanAcker/mitmproxy,Kriechi/mitmproxy,vhaupert/mitmproxy,laurmurclar/mitmproxy,ParthGanatra/mitmproxy,fimad/mitmproxy,Kriechi/mitmproxy,dufferzafar/mitmproxy,ddworken/mitmproxy,ParthGanatra/mitmproxy,jvillacorta/mitmproxy,ddworken/mitmproxy,xaxa89/mitmproxy,MatthewShao/mitmproxy
import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', - headers=None, content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) + req_patch = netlib.tutils.treq( + method='PATCH', + path=b"/path?query=param", + ) - def test_request_simple(): + + def test_curl_command(): flow = tutils.tflow(req=req_get) + result = """curl -H 'header:qvalue' 'http://address/path'""" - assert flow_export.curl_command(flow) + assert flow_export.curl_command(flow) == result flow = tutils.tflow(req=req_post) + result = """curl -X POST 'http://address/path' --data-binary 'content'""" - assert flow_export.curl_command(flow) + assert flow_export.curl_command(flow) == result + flow = tutils.tflow(req=req_patch) + result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'""" + assert flow_export.curl_command(flow) == result + +
Test exact return value of flow_export.curl_command
## Code Before: import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', headers=None, content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) def test_request_simple(): flow = tutils.tflow(req=req_get) assert flow_export.curl_command(flow) flow = tutils.tflow(req=req_post) assert flow_export.curl_command(flow) ## Instruction: Test exact return value of flow_export.curl_command ## Code After: import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) req_patch = netlib.tutils.treq( method='PATCH', path=b"/path?query=param", ) def test_curl_command(): flow = tutils.tflow(req=req_get) result = """curl -H 'header:qvalue' 'http://address/path'""" assert flow_export.curl_command(flow) == result flow = tutils.tflow(req=req_post) result = """curl -X POST 'http://address/path' --data-binary 'content'""" assert flow_export.curl_command(flow) == result flow = tutils.tflow(req=req_patch) result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'""" assert flow_export.curl_command(flow) == result
import netlib.tutils from libmproxy import flow_export from . import tutils req_get = netlib.tutils.treq( method='GET', - headers=None, content=None, ) req_post = netlib.tutils.treq( method='POST', headers=None, ) + req_patch = netlib.tutils.treq( + method='PATCH', + path=b"/path?query=param", + ) - def test_request_simple(): + + def test_curl_command(): flow = tutils.tflow(req=req_get) + result = """curl -H 'header:qvalue' 'http://address/path'""" - assert flow_export.curl_command(flow) + assert flow_export.curl_command(flow) == result ? ++++++++++ flow = tutils.tflow(req=req_post) + result = """curl -X POST 'http://address/path' --data-binary 'content'""" - assert flow_export.curl_command(flow) + assert flow_export.curl_command(flow) == result ? ++++++++++ + + flow = tutils.tflow(req=req_patch) + result = """curl -H 'header:qvalue' -X PATCH 'http://address/path?query=param' --data-binary 'content'""" + assert flow_export.curl_command(flow) == result +
ee80818b8ff12cd351581b4c1652e64561d34a4c
rest_framework_simplejwt/token_blacklist/models.py
rest_framework_simplejwt/token_blacklist/models.py
from django.contrib.auth import get_user_model from django.db import models from django.utils.six import python_2_unicode_compatible User = get_user_model() @python_2_unicode_compatible class OutstandingToken(models.Model): user = models.ForeignKey(User, on_delete=models.CASCADE) jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user)
from django.conf import settings from django.db import models from django.utils.six import python_2_unicode_compatible @python_2_unicode_compatible class OutstandingToken(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user)
Fix broken tests in 1.8-1.10
Fix broken tests in 1.8-1.10
Python
mit
davesque/django-rest-framework-simplejwt,davesque/django-rest-framework-simplejwt
- from django.contrib.auth import get_user_model + from django.conf import settings from django.db import models from django.utils.six import python_2_unicode_compatible - User = get_user_model() - - @python_2_unicode_compatible class OutstandingToken(models.Model): - user = models.ForeignKey(User, on_delete=models.CASCADE) + user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user)
Fix broken tests in 1.8-1.10
## Code Before: from django.contrib.auth import get_user_model from django.db import models from django.utils.six import python_2_unicode_compatible User = get_user_model() @python_2_unicode_compatible class OutstandingToken(models.Model): user = models.ForeignKey(User, on_delete=models.CASCADE) jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user) ## Instruction: Fix broken tests in 1.8-1.10 ## Code After: from django.conf import settings from django.db import models from django.utils.six import python_2_unicode_compatible @python_2_unicode_compatible class OutstandingToken(models.Model): user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user)
- from django.contrib.auth import get_user_model + from django.conf import settings from django.db import models from django.utils.six import python_2_unicode_compatible - User = get_user_model() - - @python_2_unicode_compatible class OutstandingToken(models.Model): - user = models.ForeignKey(User, on_delete=models.CASCADE) ? - ^ + user = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE) ? ^^^^^^^^^^^^^^^^^^^^^^ jti = models.UUIDField(unique=True) token = models.TextField() created_at = models.DateTimeField() expires_at = models.DateTimeField() class Meta: ordering = ('user',) def __str__(self): return 'Token for {} ({})'.format( self.user, self.jti, ) @python_2_unicode_compatible class BlacklistedToken(models.Model): token = models.OneToOneField(OutstandingToken, on_delete=models.CASCADE) blacklisted_at = models.DateTimeField(auto_now_add=True) def __str__(self): return 'Blacklisted token for {}'.format(self.token.user)
5ac8e4619473275f2f0b26b8a9b64049d793a4ed
rmqid/__init__.py
rmqid/__init__.py
__version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish
__version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger().addHandler(NullHandler())
Add a NullHandler so logging warnings are not emitted if no logger is setup
Add a NullHandler so logging warnings are not emitted if no logger is setup
Python
bsd-3-clause
jonahbull/rabbitpy,gmr/rabbitpy,gmr/rabbitpy
__version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish + import logging + + try: + from logging import NullHandler + except ImportError: + # Python 2.6 does not have a NullHandler + class NullHandler(logging.Handler): + def emit(self, record): + pass + + logging.getLogger().addHandler(NullHandler()) +
Add a NullHandler so logging warnings are not emitted if no logger is setup
## Code Before: __version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish ## Instruction: Add a NullHandler so logging warnings are not emitted if no logger is setup ## Code After: __version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish import logging try: from logging import NullHandler except ImportError: # Python 2.6 does not have a NullHandler class NullHandler(logging.Handler): def emit(self, record): pass logging.getLogger().addHandler(NullHandler())
__version__ = '0.3.0' from rmqid.connection import Connection from rmqid.exchange import Exchange from rmqid.message import Message from rmqid.queue import Queue from rmqid.tx import Tx from rmqid.simple import consumer from rmqid.simple import get from rmqid.simple import publish + + import logging + + try: + from logging import NullHandler + except ImportError: + # Python 2.6 does not have a NullHandler + class NullHandler(logging.Handler): + def emit(self, record): + pass + + logging.getLogger().addHandler(NullHandler())
257e8d2e6d1dc3c10eb7fc26c3deacaf4133bd9b
enactiveagents/view/agentevents.py
enactiveagents/view/agentevents.py
import events class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): if event.agent not in self.preparation_history: self.preparation_history[event.agent] = [] self.preparation_history[event.agent].append(event.action) elif isinstance(event, events.AgentEnactionEvent): if event.agent not in self.enaction_history: self.enaction_history[event.agent] = [] self.enaction_history[event.agent].append(event.action) elif isinstance(event, events.TickEvent): pass
import events import json class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): if str(event.agent) not in self.preparation_history: self.preparation_history[str(event.agent)] = [] self.preparation_history[str(event.agent)].append(str(event.action)) if len(self.preparation_history) > 20: self.preparation_history.pop(0) elif isinstance(event, events.AgentEnactionEvent): if str(event.agent) not in self.enaction_history: self.enaction_history[str(event.agent)] = [] self.enaction_history[str(event.agent)].append(str(event.action)) if len(self.enaction_history) > 20: self.enaction_history.pop(0) elif isinstance(event, events.TickEvent): self.write_to_file() def write_to_file(self): """ Write the history to the traces file. """ d = dict() d["preparation_history"] = self.preparation_history d["enaction_history"] = self.enaction_history with open(self.file_path,'w+') as f: json.dump(d, f)
Write agent events to a traces history file for the website.
Write agent events to a traces history file for the website.
Python
mit
Beskhue/enactive-agents,Beskhue/enactive-agents,Beskhue/enactive-agents
import events + import json class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): - if event.agent not in self.preparation_history: + if str(event.agent) not in self.preparation_history: - self.preparation_history[event.agent] = [] + self.preparation_history[str(event.agent)] = [] - self.preparation_history[event.agent].append(event.action) + self.preparation_history[str(event.agent)].append(str(event.action)) + + if len(self.preparation_history) > 20: + self.preparation_history.pop(0) elif isinstance(event, events.AgentEnactionEvent): - if event.agent not in self.enaction_history: + if str(event.agent) not in self.enaction_history: - self.enaction_history[event.agent] = [] + self.enaction_history[str(event.agent)] = [] - self.enaction_history[event.agent].append(event.action) + self.enaction_history[str(event.agent)].append(str(event.action)) + + if len(self.enaction_history) > 20: + self.enaction_history.pop(0) elif isinstance(event, events.TickEvent): - pass + self.write_to_file() + + def write_to_file(self): + """ + Write the history to the traces file. + """ + + d = dict() + d["preparation_history"] = self.preparation_history + d["enaction_history"] = self.enaction_history + with open(self.file_path,'w+') as f: + json.dump(d, f)
Write agent events to a traces history file for the website.
## Code Before: import events class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): if event.agent not in self.preparation_history: self.preparation_history[event.agent] = [] self.preparation_history[event.agent].append(event.action) elif isinstance(event, events.AgentEnactionEvent): if event.agent not in self.enaction_history: self.enaction_history[event.agent] = [] self.enaction_history[event.agent].append(event.action) elif isinstance(event, events.TickEvent): pass ## Instruction: Write agent events to a traces history file for the website. ## Code After: import events import json class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): if str(event.agent) not in self.preparation_history: self.preparation_history[str(event.agent)] = [] self.preparation_history[str(event.agent)].append(str(event.action)) if len(self.preparation_history) > 20: self.preparation_history.pop(0) elif isinstance(event, events.AgentEnactionEvent): if str(event.agent) not in self.enaction_history: self.enaction_history[str(event.agent)] = [] self.enaction_history[str(event.agent)].append(str(event.action)) if len(self.enaction_history) > 20: self.enaction_history.pop(0) elif isinstance(event, events.TickEvent): self.write_to_file() def write_to_file(self): """ Write the history to the traces file. """ d = dict() d["preparation_history"] = self.preparation_history d["enaction_history"] = self.enaction_history with open(self.file_path,'w+') as f: json.dump(d, f)
import events + import json class AgentEvents(events.EventListener): """ View class """ def __init__(self, file_path): """ :param file_path: The path of the file to output the history to. """ self.file_path = file_path self.preparation_history = dict() self.enaction_history = dict() def notify(self, event): if isinstance(event, events.AgentPreparationEvent): - if event.agent not in self.preparation_history: + if str(event.agent) not in self.preparation_history: ? ++++ + - self.preparation_history[event.agent] = [] + self.preparation_history[str(event.agent)] = [] ? ++++ + - self.preparation_history[event.agent].append(event.action) + self.preparation_history[str(event.agent)].append(str(event.action)) ? ++++ + ++++ + + + if len(self.preparation_history) > 20: + self.preparation_history.pop(0) elif isinstance(event, events.AgentEnactionEvent): - if event.agent not in self.enaction_history: + if str(event.agent) not in self.enaction_history: ? ++++ + - self.enaction_history[event.agent] = [] + self.enaction_history[str(event.agent)] = [] ? ++++ + - self.enaction_history[event.agent].append(event.action) + self.enaction_history[str(event.agent)].append(str(event.action)) ? ++++ + ++++ + + + if len(self.enaction_history) > 20: + self.enaction_history.pop(0) elif isinstance(event, events.TickEvent): - pass + self.write_to_file() + + def write_to_file(self): + """ + Write the history to the traces file. + """ + + d = dict() + d["preparation_history"] = self.preparation_history + d["enaction_history"] = self.enaction_history + with open(self.file_path,'w+') as f: + json.dump(d, f)
a24d6a25cb7ee5101e8131a9719744f79b23c11b
examples/quotes/quotes.py
examples/quotes/quotes.py
import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) quote = random.choice(quotes) nw0.send_notification(address, "quote", quote) time.sleep(0.5) if __name__ == '__main__': main(*sys.argv[1:])
import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) quote = random.choice(quotes) nw0.send_notification(my_address, "quote", quote) time.sleep(1) if __name__ == '__main__': main(*sys.argv[1:])
Send notification to the correct address
Send notification to the correct address
Python
mit
tjguk/networkzero,tjguk/networkzero,tjguk/networkzero
import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] - + for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) - quote = random.choice(quotes) - nw0.send_notification(address, "quote", quote) + quote = random.choice(quotes) + nw0.send_notification(my_address, "quote", quote) + - time.sleep(0.5) + time.sleep(1) if __name__ == '__main__': main(*sys.argv[1:])
Send notification to the correct address
## Code Before: import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) quote = random.choice(quotes) nw0.send_notification(address, "quote", quote) time.sleep(0.5) if __name__ == '__main__': main(*sys.argv[1:]) ## Instruction: Send notification to the correct address ## Code After: import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) quote = random.choice(quotes) nw0.send_notification(my_address, "quote", quote) time.sleep(1) if __name__ == '__main__': main(*sys.argv[1:])
import sys print(sys.version_info) import random import time import networkzero as nw0 quotes = [ "Humpty Dumpty sat on a wall", "Hickory Dickory Dock", "Baa Baa Black Sheep", "Old King Cole was a merry old sould", ] def main(address_pattern=None): my_name = input("Name: ") my_address = nw0.advertise(my_name, address_pattern) print("Advertising %s on %s" % (my_name, my_address)) while True: services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name] - + for name, address in services: topic, message = nw0.wait_for_notification(address, "quote", wait_for_s=0) if topic: print("%s says: %s" % (name, message)) - quote = random.choice(quotes) - nw0.send_notification(address, "quote", quote) + quote = random.choice(quotes) + nw0.send_notification(my_address, "quote", quote) + - time.sleep(0.5) ? ^^^ + time.sleep(1) ? ^ if __name__ == '__main__': main(*sys.argv[1:])
ffde5305a2182e566384887d51e4fde90adc9908
runtests.py
runtests.py
import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() failures = test_runner.run_tests(["tests"]) sys.exit(bool(failures))
import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": tests = "tests" if len(sys.argv) == 1 else sys.argv[1] os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() failures = test_runner.run_tests([tests]) sys.exit(bool(failures))
Make it possible to run individual tests.
Tests: Make it possible to run individual tests.
Python
agpl-3.0
etesync/journal-manager
import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": + tests = "tests" if len(sys.argv) == 1 else sys.argv[1] os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() - failures = test_runner.run_tests(["tests"]) + failures = test_runner.run_tests([tests]) sys.exit(bool(failures))
Make it possible to run individual tests.
## Code Before: import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() failures = test_runner.run_tests(["tests"]) sys.exit(bool(failures)) ## Instruction: Make it possible to run individual tests. ## Code After: import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": tests = "tests" if len(sys.argv) == 1 else sys.argv[1] os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() failures = test_runner.run_tests([tests]) sys.exit(bool(failures))
import os import sys import django from django.conf import settings from django.test.utils import get_runner if __name__ == "__main__": + tests = "tests" if len(sys.argv) == 1 else sys.argv[1] os.environ['DJANGO_SETTINGS_MODULE'] = 'tests.test_settings' django.setup() TestRunner = get_runner(settings) test_runner = TestRunner() - failures = test_runner.run_tests(["tests"]) ? - - + failures = test_runner.run_tests([tests]) sys.exit(bool(failures))
09d356f7b124368ac2ca80efa981d115ea847196
django_ethereum_events/web3_service.py
django_ethereum_events/web3_service.py
from django.conf import settings from web3 import Web3, RPCProvider from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: rpc_provider = RPCProvider( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, ssl=settings.ETHEREUM_NODE_SSL, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__()
from django.conf import settings from web3 import Web3 try: from web3 import HTTPProvider RPCProvider = None except ImportError: from web3 import RPCProvider HTTPProvider = None from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: if HTTPProvider is not None: uri = "{scheme}://{host}:{port}".format( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, scheme="https" if settings.ETHEREUM_NODE_SSL else "http", ) rpc_provider = HTTPProvider( endpoint_uri=uri, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) elif RPCProvider is not None: rpc_provider = RPCProvider( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, ssl=settings.ETHEREUM_NODE_SSL, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) else: raise ValueError("Cannot instantiate any RPC provider") self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__()
Support for Web3 4.0beta: HTTPProvider
Support for Web3 4.0beta: HTTPProvider In Web3 3.16 the class is called RPCProvider, but in the upcoming 4.0 series it's replaced with HTTPProvider. This commit ensures both versions are supported in this regard.
Python
mit
artemistomaras/django-ethereum-events,artemistomaras/django-ethereum-events
from django.conf import settings + from web3 import Web3 + try: + from web3 import HTTPProvider + RPCProvider = None + except ImportError: - from web3 import Web3, RPCProvider + from web3 import RPCProvider + HTTPProvider = None from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: - rpc_provider = RPCProvider( + if HTTPProvider is not None: + uri = "{scheme}://{host}:{port}".format( - host=settings.ETHEREUM_NODE_HOST, + host=settings.ETHEREUM_NODE_HOST, - port=settings.ETHEREUM_NODE_PORT, + port=settings.ETHEREUM_NODE_PORT, - ssl=settings.ETHEREUM_NODE_SSL, + scheme="https" if settings.ETHEREUM_NODE_SSL else "http", + ) + rpc_provider = HTTPProvider( + endpoint_uri=uri, - timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) + timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) - ) + ) + elif RPCProvider is not None: + rpc_provider = RPCProvider( + host=settings.ETHEREUM_NODE_HOST, + port=settings.ETHEREUM_NODE_PORT, + ssl=settings.ETHEREUM_NODE_SSL, + timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) + ) + else: + raise ValueError("Cannot instantiate any RPC provider") self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__()
Support for Web3 4.0beta: HTTPProvider
## Code Before: from django.conf import settings from web3 import Web3, RPCProvider from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: rpc_provider = RPCProvider( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, ssl=settings.ETHEREUM_NODE_SSL, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__() ## Instruction: Support for Web3 4.0beta: HTTPProvider ## Code After: from django.conf import settings from web3 import Web3 try: from web3 import HTTPProvider RPCProvider = None except ImportError: from web3 import RPCProvider HTTPProvider = None from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: if HTTPProvider is not None: uri = "{scheme}://{host}:{port}".format( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, scheme="https" if settings.ETHEREUM_NODE_SSL else "http", ) rpc_provider = HTTPProvider( endpoint_uri=uri, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) elif RPCProvider is not None: rpc_provider = RPCProvider( host=settings.ETHEREUM_NODE_HOST, port=settings.ETHEREUM_NODE_PORT, ssl=settings.ETHEREUM_NODE_SSL, timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ) else: raise ValueError("Cannot instantiate any RPC provider") self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__()
from django.conf import settings + from web3 import Web3 + try: + from web3 import HTTPProvider + RPCProvider = None + except ImportError: - from web3 import Web3, RPCProvider ? ------ + from web3 import RPCProvider ? ++++ + HTTPProvider = None from .singleton import Singleton class Web3Service(metaclass=Singleton): """Creates a `web3` instance based on the given `RPCProvider`.""" def __init__(self, *args, **kwargs): """Initializes the `web3` object. Args: rpc_provider (:obj:`Provider`, optional): Valid `web3` Provider instance. """ rpc_provider = kwargs.pop('rpc_provider', None) if not rpc_provider: - rpc_provider = RPCProvider( + if HTTPProvider is not None: + uri = "{scheme}://{host}:{port}".format( - host=settings.ETHEREUM_NODE_HOST, + host=settings.ETHEREUM_NODE_HOST, ? ++++ - port=settings.ETHEREUM_NODE_PORT, + port=settings.ETHEREUM_NODE_PORT, ? ++++ - ssl=settings.ETHEREUM_NODE_SSL, + scheme="https" if settings.ETHEREUM_NODE_SSL else "http", + ) + rpc_provider = HTTPProvider( + endpoint_uri=uri, - timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) + timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) ? ++++ - ) + ) ? ++++ + elif RPCProvider is not None: + rpc_provider = RPCProvider( + host=settings.ETHEREUM_NODE_HOST, + port=settings.ETHEREUM_NODE_PORT, + ssl=settings.ETHEREUM_NODE_SSL, + timeout=getattr(settings, "ETHEREUM_NODE_TIMEOUT", 10) + ) + else: + raise ValueError("Cannot instantiate any RPC provider") self.web3 = Web3(rpc_provider) super(Web3Service, self).__init__()
d3cb08d45af60aaf06757ad230a2a33bc3615543
apps/organizations/middleware.py
apps/organizations/middleware.py
from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404
from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): if request.subdomain is None: return try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404
Remove subdomain check on pages where subdomain is none
Remove subdomain check on pages where subdomain is none
Python
mit
xobb1t/ddash2013,xobb1t/ddash2013
from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): + if request.subdomain is None: + return try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404
Remove subdomain check on pages where subdomain is none
## Code Before: from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404 ## Instruction: Remove subdomain check on pages where subdomain is none ## Code After: from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): if request.subdomain is None: return try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404
from django.http import Http404 from .models import Organization class OrganizationMiddleware(object): def process_request(self, request): + if request.subdomain is None: + return try: request.organization = Organization.objects.get( slug__iexact=request.subdomain ) except Organization.DoesNotExist: raise Http404
b7a0653cdb2c20def38a687963763b75455ebbcb
conftest.py
conftest.py
from __future__ import absolute_import, division, print_function from dials.conftest import regression_data, run_in_tmpdir
from __future__ import absolute_import, division, print_function from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
Add --regression command line option
Add --regression command line option
Python
bsd-3-clause
xia2/i19
from __future__ import absolute_import, division, print_function - from dials.conftest import regression_data, run_in_tmpdir + from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
Add --regression command line option
## Code Before: from __future__ import absolute_import, division, print_function from dials.conftest import regression_data, run_in_tmpdir ## Instruction: Add --regression command line option ## Code After: from __future__ import absolute_import, division, print_function from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir
from __future__ import absolute_import, division, print_function - from dials.conftest import regression_data, run_in_tmpdir + from dials.conftest import pytest_addoption, regression_data, run_in_tmpdir ? ++++++++++++++++++
7c3cf9e430bee4451e817ccc3d32884ed0c5f8e9
bakeit/uploader.py
bakeit/uploader.py
try: from urllib.request import urlopen, Request, HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: response = json.loads(e.read()) raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"]
try: from urllib.request import urlopen, Request from urllib.error import HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: response = json.loads(e.read().decode("utf8")) raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"]
Fix Python3 error when decoding the response.
fix: Fix Python3 error when decoding the response.
Python
mit
skorokithakis/bakeit
try: - from urllib.request import urlopen, Request, HTTPError + from urllib.request import urlopen, Request + from urllib.error import HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: - response = json.loads(e.read()) + response = json.loads(e.read().decode("utf8")) raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"]
Fix Python3 error when decoding the response.
## Code Before: try: from urllib.request import urlopen, Request, HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: response = json.loads(e.read()) raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"] ## Instruction: Fix Python3 error when decoding the response. ## Code After: try: from urllib.request import urlopen, Request from urllib.error import HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: response = json.loads(e.read().decode("utf8")) raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"]
try: - from urllib.request import urlopen, Request, HTTPError ? ----------- + from urllib.request import urlopen, Request + from urllib.error import HTTPError except ImportError: from urllib2 import urlopen, Request, HTTPError import json class PasteryUploader(): def __init__(self, api_key): """ Initialize an Uploader instance with the given API key. """ self.api_key = api_key def upload(self, body, title="", language=None, duration=None, max_views=0): """ Upload the given body with the specified language type. """ url = "https://www.pastery.net/api/paste/?api_key=%s" % self.api_key if title: url += "&title=%s" % title if language: url += "&language=%s" % language if duration: url += "&duration=%s" % duration if max_views: url += "&max_views=%s" % max_views body = bytes(body.encode("utf8")) req = Request(url, data=body, headers={'User-Agent': u'Mozilla/5.0 (Python) bakeit library'}) try: response = urlopen(req) except HTTPError as e: - response = json.loads(e.read()) + response = json.loads(e.read().decode("utf8")) ? ++++++++++++++ + raise RuntimeError(response["error_msg"]) response = json.loads(response.read().decode("utf8")) return response["url"]
10948cd88d51383e13af0a116703984752092c6a
jenkinsapi_tests/systests/test_jenkins_matrix.py
jenkinsapi_tests/systests/test_jenkins_matrix.py
''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) b = job.get_last_build() while b.is_running(): time.sleep(1) s = set() for r in b.get_matrix_runs(): self.assertEquals(r.get_number(), b.get_number()) self.assertEquals(r.get_upstream_build(), b) m = re.search(u'\xbb (.*) #\\d+$', r.name) self.assertIsNotNone(m) s.add(m.group(1)) # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. self.assertFalse(b.get_master_job_name()) self.assertEqual(s, set(['one', 'two', 'three'])) if __name__ == '__main__': unittest.main()
''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) build = job.get_last_build() while build.is_running(): time.sleep(1) set_of_groups = set() for run in build.get_matrix_runs(): self.assertEquals(run.get_number(), build.get_number()) self.assertEquals(run.get_upstream_build(), build) match_result = re.search(u'\xbb (.*) #\\d+$', run.name) self.assertIsNotNone(match_result) set_of_groups.add(match_result.group(1)) build.get_master_job_name() # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. self.assertFalse(build.get_master_job_name()) self.assertEqual(set_of_groups, set(['one', 'two', 'three'])) if __name__ == '__main__': unittest.main()
Tidy up this test - still quite bad & useless.
Tidy up this test - still quite bad & useless.
Python
mit
imsardine/jenkinsapi,salimfadhley/jenkinsapi,JohnLZeller/jenkinsapi,JohnLZeller/jenkinsapi,aerickson/jenkinsapi,domenkozar/jenkinsapi,zaro0508/jenkinsapi,imsardine/jenkinsapi,zaro0508/jenkinsapi,jduan/jenkinsapi,mistermocha/jenkinsapi,domenkozar/jenkinsapi,salimfadhley/jenkinsapi,zaro0508/jenkinsapi,mistermocha/jenkinsapi,aerickson/jenkinsapi,jduan/jenkinsapi,JohnLZeller/jenkinsapi,mistermocha/jenkinsapi,imsardine/jenkinsapi
''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) - b = job.get_last_build() + build = job.get_last_build() - while b.is_running(): + while build.is_running(): time.sleep(1) - s = set() + set_of_groups = set() - for r in b.get_matrix_runs(): + for run in build.get_matrix_runs(): - self.assertEquals(r.get_number(), b.get_number()) + self.assertEquals(run.get_number(), build.get_number()) - self.assertEquals(r.get_upstream_build(), b) + self.assertEquals(run.get_upstream_build(), build) - m = re.search(u'\xbb (.*) #\\d+$', r.name) + match_result = re.search(u'\xbb (.*) #\\d+$', run.name) - self.assertIsNotNone(m) + self.assertIsNotNone(match_result) - s.add(m.group(1)) + set_of_groups.add(match_result.group(1)) + build.get_master_job_name() # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. - self.assertFalse(b.get_master_job_name()) + self.assertFalse(build.get_master_job_name()) - self.assertEqual(s, set(['one', 'two', 'three'])) + self.assertEqual(set_of_groups, set(['one', 'two', 'three'])) if __name__ == '__main__': unittest.main()
Tidy up this test - still quite bad & useless.
## Code Before: ''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) b = job.get_last_build() while b.is_running(): time.sleep(1) s = set() for r in b.get_matrix_runs(): self.assertEquals(r.get_number(), b.get_number()) self.assertEquals(r.get_upstream_build(), b) m = re.search(u'\xbb (.*) #\\d+$', r.name) self.assertIsNotNone(m) s.add(m.group(1)) # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. self.assertFalse(b.get_master_job_name()) self.assertEqual(s, set(['one', 'two', 'three'])) if __name__ == '__main__': unittest.main() ## Instruction: Tidy up this test - still quite bad & useless. ## Code After: ''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) build = job.get_last_build() while build.is_running(): time.sleep(1) set_of_groups = set() for run in build.get_matrix_runs(): self.assertEquals(run.get_number(), build.get_number()) self.assertEquals(run.get_upstream_build(), build) match_result = re.search(u'\xbb (.*) #\\d+$', run.name) self.assertIsNotNone(match_result) set_of_groups.add(match_result.group(1)) build.get_master_job_name() # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. self.assertFalse(build.get_master_job_name()) self.assertEqual(set_of_groups, set(['one', 'two', 'three'])) if __name__ == '__main__': unittest.main()
''' System tests for `jenkinsapi.jenkins` module. ''' import re import time import unittest from jenkinsapi_tests.systests.base import BaseSystemTest from jenkinsapi_tests.systests.job_configs import MATRIX_JOB from jenkinsapi_tests.test_utils.random_strings import random_string class TestMatrixJob(BaseSystemTest): def test_invoke_matrix_job(self): job_name = 'create_%s' % random_string() job = self.jenkins.create_job(job_name, MATRIX_JOB) job.invoke(block=True) - b = job.get_last_build() + build = job.get_last_build() ? ++++ - while b.is_running(): + while build.is_running(): ? ++++ time.sleep(1) - s = set() + set_of_groups = set() - for r in b.get_matrix_runs(): + for run in build.get_matrix_runs(): ? ++ ++++ - self.assertEquals(r.get_number(), b.get_number()) + self.assertEquals(run.get_number(), build.get_number()) ? ++ ++++ - self.assertEquals(r.get_upstream_build(), b) + self.assertEquals(run.get_upstream_build(), build) ? ++ ++++ - m = re.search(u'\xbb (.*) #\\d+$', r.name) + match_result = re.search(u'\xbb (.*) #\\d+$', run.name) ? +++++++++++ ++ - self.assertIsNotNone(m) + self.assertIsNotNone(match_result) ? +++++++++++ - s.add(m.group(1)) + set_of_groups.add(match_result.group(1)) + build.get_master_job_name() # This is a bad test, it simply verifies that this function does # not crash on a build from a matrix job. - self.assertFalse(b.get_master_job_name()) + self.assertFalse(build.get_master_job_name()) ? ++++ - self.assertEqual(s, set(['one', 'two', 'three'])) + self.assertEqual(set_of_groups, set(['one', 'two', 'three'])) ? ++++++++++++ if __name__ == '__main__': unittest.main()
86dca4a7d3c1574af9da85e5a2f10b84d18d28c0
blueprints/aws_backup_plans/delete.py
blueprints/aws_backup_plans/delete.py
from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) backup_plan_name=resource.name backup_vault_name=backup_plan_name+'backup-vault' set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) try: set_progress("Deleting the backup plan vault...") client.delete_backup_vault( BackupVaultName=backup_vault_name) set_progress("Deleting the backup plan...") client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", ""
from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) set_progress("Deleting the backup plan...") try: client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", ""
Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working"
Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working"
Python
apache-2.0
CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge,CloudBoltSoftware/cloudbolt-forge
from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) + - backup_plan_name=resource.name - backup_vault_name=backup_plan_name+'backup-vault' set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) - + set_progress("Deleting the backup plan...") try: - set_progress("Deleting the backup plan vault...") - client.delete_backup_vault( - BackupVaultName=backup_vault_name) - - set_progress("Deleting the backup plan...") client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", "" -
Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working"
## Code Before: from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) backup_plan_name=resource.name backup_vault_name=backup_plan_name+'backup-vault' set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) try: set_progress("Deleting the backup plan vault...") client.delete_backup_vault( BackupVaultName=backup_vault_name) set_progress("Deleting the backup plan...") client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", "" ## Instruction: Revert "[Dev-20546] AwSBackPlan-Blueprint is broken-Teardown is not working" ## Code After: from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) set_progress("Deleting the backup plan...") try: client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", ""
from common.methods import set_progress from azure.common.credentials import ServicePrincipalCredentials from botocore.exceptions import ClientError from resourcehandlers.aws.models import AWSHandler import boto3 def run(job, **kwargs): resource = kwargs.pop('resources').first() backup_plan_id = resource.attributes.get(field__name='backup_plan_id').value rh_id = resource.attributes.get(field__name='aws_rh_id').value region = resource.attributes.get(field__name='aws_region').value rh = AWSHandler.objects.get(id=rh_id) + - backup_plan_name=resource.name - backup_vault_name=backup_plan_name+'backup-vault' set_progress("Connecting to aws backups...") client = boto3.client('backup', region_name=region, aws_access_key_id=rh.serviceaccount, aws_secret_access_key=rh.servicepasswd ) - + set_progress("Deleting the backup plan...") try: - set_progress("Deleting the backup plan vault...") - client.delete_backup_vault( - BackupVaultName=backup_vault_name) - - set_progress("Deleting the backup plan...") client.delete_backup_plan(BackupPlanId=backup_plan_id) except Exception as e: return "FAILURE", "Backup plan could not be deleted", e return "SUCCESS", "The network security group has been succesfully deleted", ""
c9ecacdb04f3f8df4f85057ad0d3c69df9481122
core/utils/check_sanity.py
core/utils/check_sanity.py
import os from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True
import os from core.utils.Executor import _convert_subprocess_cmd import subprocess from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True def check_mirage(): try: subprocess.check_call( _convert_subprocess_cmd('which mirage') ) except subprocess.CalledProcessError: return False else: return True
Check if mirage is installed
Check if mirage is installed
Python
apache-2.0
onyb/dune,adyasha/dune,adyasha/dune,adyasha/dune
import os + from core.utils.Executor import _convert_subprocess_cmd + import subprocess from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True + + def check_mirage(): + try: + subprocess.check_call( + _convert_subprocess_cmd('which mirage') + ) + except subprocess.CalledProcessError: + return False + else: + return True +
Check if mirage is installed
## Code Before: import os from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True ## Instruction: Check if mirage is installed ## Code After: import os from core.utils.Executor import _convert_subprocess_cmd import subprocess from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True def check_mirage(): try: subprocess.check_call( _convert_subprocess_cmd('which mirage') ) except subprocess.CalledProcessError: return False else: return True
import os + from core.utils.Executor import _convert_subprocess_cmd + import subprocess from core.exceptions.Exceptions import OPAMConfigurationExeception def check_environment() -> bool: __opam_env__ = [ 'CAML_LD_LIBRARY_PATH', 'MANPATH', 'PERL5LIB', 'OCAML_TOPLEVEL_PATH', 'PATH' ] for var in __opam_env__: if not os.environ.get(var, None): raise OPAMConfigurationExeception PATH = os.environ.get('PATH') for path in PATH.split(':'): if path.endswith( os.path.join('.opam', 'system', 'bin') ): return True + + + def check_mirage(): + try: + subprocess.check_call( + _convert_subprocess_cmd('which mirage') + ) + except subprocess.CalledProcessError: + return False + else: + return True
e3b0ccb529dca19bb3882f9caad82dbd965c9ae0
onnx/__init__.py
onnx/__init__.py
from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params Takes a file-like object (has to implement fileno that returns a file descriptor) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() if isinstance(obj, str) or (sys.version_info[0] == 2 and isinstance(obj, unicode_literals.unicode_or_str)): with open(obj, 'rb') as f: model.ParseFromString(f.read()) else: model.ParseFromString(obj.read()) return model
from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params Takes a file-like object (has "read" function) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() if hasattr(obj, 'read') and callable(obj.read): model.ParseFromString(obj.read()) else: with open(obj, 'rb') as f: model.ParseFromString(f.read()) return model
Fix string/file-like object detection in onnx.load
Fix string/file-like object detection in onnx.load
Python
apache-2.0
onnx/onnx,onnx/onnx,onnx/onnx,onnx/onnx
from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params - Takes a file-like object (has to implement fileno that returns a file descriptor) + Takes a file-like object (has "read" function) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() - if isinstance(obj, str) or (sys.version_info[0] == 2 and - isinstance(obj, unicode_literals.unicode_or_str)): + if hasattr(obj, 'read') and callable(obj.read): + model.ParseFromString(obj.read()) + else: with open(obj, 'rb') as f: model.ParseFromString(f.read()) - else: - model.ParseFromString(obj.read()) return model
Fix string/file-like object detection in onnx.load
## Code Before: from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params Takes a file-like object (has to implement fileno that returns a file descriptor) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() if isinstance(obj, str) or (sys.version_info[0] == 2 and isinstance(obj, unicode_literals.unicode_or_str)): with open(obj, 'rb') as f: model.ParseFromString(f.read()) else: model.ParseFromString(obj.read()) return model ## Instruction: Fix string/file-like object detection in onnx.load ## Code After: from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params Takes a file-like object (has "read" function) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() if hasattr(obj, 'read') and callable(obj.read): model.ParseFromString(obj.read()) else: with open(obj, 'rb') as f: model.ParseFromString(f.read()) return model
from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals from .onnx_pb2 import * import sys def load(obj): ''' Loads a binary protobuf that stores onnx graph @params - Takes a file-like object (has to implement fileno that returns a file descriptor) + Takes a file-like object (has "read" function) or a string containing a file name @return ONNX ModelProto object ''' model = ModelProto() - if isinstance(obj, str) or (sys.version_info[0] == 2 and - isinstance(obj, unicode_literals.unicode_or_str)): + if hasattr(obj, 'read') and callable(obj.read): + model.ParseFromString(obj.read()) + else: with open(obj, 'rb') as f: model.ParseFromString(f.read()) - else: - model.ParseFromString(obj.read()) return model
18818a8dfebcc44f9e8b582c15d6185f9a7a0c45
minicms/templatetags/cms.py
minicms/templatetags/cms.py
from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active if request.path != '/': active = None active_len = 0 # Normalize path path = request.path.rstrip('/') + '/' for item in menu: # Normalize path url = item['url'].rstrip('/') + '/' if path.startswith(url) and len(url) > active_len: active = item active_len = len(url) if active is not None: active['active'] = True return {'menu': menu}
from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active active = None active_len = 0 # Normalize path path = request.path.rstrip('/') + '/' for item in menu: # Normalize path url = item['url'].rstrip('/') + '/' # Root is only active if you have a "Home" link if path != '/' and url == '/': continue if path.startswith(url) and len(url) > active_len: active = item active_len = len(url) if active is not None: active['active'] = True return {'menu': menu}
Allow "Home" to be active menu item
Allow "Home" to be active menu item
Python
bsd-3-clause
adieu/allbuttonspressed,adieu/allbuttonspressed
from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active - if request.path != '/': - active = None + active = None - active_len = 0 + active_len = 0 + # Normalize path + path = request.path.rstrip('/') + '/' + for item in menu: # Normalize path - path = request.path.rstrip('/') + '/' - for item in menu: - # Normalize path - url = item['url'].rstrip('/') + '/' + url = item['url'].rstrip('/') + '/' + # Root is only active if you have a "Home" link + if path != '/' and url == '/': + continue - if path.startswith(url) and len(url) > active_len: + if path.startswith(url) and len(url) > active_len: - active = item + active = item - active_len = len(url) + active_len = len(url) - if active is not None: + if active is not None: - active['active'] = True + active['active'] = True return {'menu': menu}
Allow "Home" to be active menu item
## Code Before: from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active if request.path != '/': active = None active_len = 0 # Normalize path path = request.path.rstrip('/') + '/' for item in menu: # Normalize path url = item['url'].rstrip('/') + '/' if path.startswith(url) and len(url) > active_len: active = item active_len = len(url) if active is not None: active['active'] = True return {'menu': menu} ## Instruction: Allow "Home" to be active menu item ## Code After: from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active active = None active_len = 0 # Normalize path path = request.path.rstrip('/') + '/' for item in menu: # Normalize path url = item['url'].rstrip('/') + '/' # Root is only active if you have a "Home" link if path != '/' and url == '/': continue if path.startswith(url) and len(url) > active_len: active = item active_len = len(url) if active is not None: active['active'] = True return {'menu': menu}
from ..models import Block from django.template import Library register = Library() @register.simple_tag def show_block(name): try: return Block.objects.get(name=name).content except Block.DoesNotExist: return '' except Block.MultipleObjectsReturned: return 'Error: Multiple blocks for "%s"' % name @register.inclusion_tag('minicms/menu.html', takes_context=True) def show_menu(context, name='menu'): request = context['request'] menu = [] try: for line in Block.objects.get(name=name).content.splitlines(): line = line.rstrip() try: title, url = line.rsplit(' ', 1) except: continue menu.append({'title': title.strip(), 'url': url}) except Block.DoesNotExist: pass # Mark the best-matching URL as active - if request.path != '/': - active = None ? ---- + active = None - active_len = 0 ? ---- + active_len = 0 + # Normalize path + path = request.path.rstrip('/') + '/' + for item in menu: # Normalize path - path = request.path.rstrip('/') + '/' - for item in menu: - # Normalize path - url = item['url'].rstrip('/') + '/' ? ---- + url = item['url'].rstrip('/') + '/' + # Root is only active if you have a "Home" link + if path != '/' and url == '/': + continue - if path.startswith(url) and len(url) > active_len: ? ---- + if path.startswith(url) and len(url) > active_len: - active = item ? ---- + active = item - active_len = len(url) ? ---- + active_len = len(url) - if active is not None: ? ---- + if active is not None: - active['active'] = True ? ---- + active['active'] = True return {'menu': menu}
eb33d70bfda4857fbd76616cf3bf7fb7d7feec71
spoj/00005/palin.py
spoj/00005/palin.py
def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) for j in range(i + 1, mid): palin[j] = '0' for j in range(mid, n): mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k))
def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) for j in range(i + 1, (n + 1) // 2): palin[j] = '0' for j in range((n + 1) // 2, n): mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k))
Fix bug in ranges (to middle)
Fix bug in ranges (to middle) - in SPOJ palin Signed-off-by: Karel Ha <70f8965fdfb04f1fc0e708a55d9e822c449f57d3@gmail.com>
Python
mit
mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming,mathemage/CompetitiveProgramming
def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) - for j in range(i + 1, mid): + for j in range(i + 1, (n + 1) // 2): palin[j] = '0' - for j in range(mid, n): + for j in range((n + 1) // 2, n): mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k))
Fix bug in ranges (to middle)
## Code Before: def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) for j in range(i + 1, mid): palin[j] = '0' for j in range(mid, n): mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k)) ## Instruction: Fix bug in ranges (to middle) ## Code After: def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) for j in range(i + 1, (n + 1) // 2): palin[j] = '0' for j in range((n + 1) // 2, n): mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k))
def next_palindrome(k): palin = list(k) n = len(k) mid = n // 2 # case 1: forward right just_copy = False for i in range(mid, n): mirrored = n - 1 - i if k[i] < k[mirrored]: just_copy = True if just_copy: palin[i] = palin[mirrored] # case 2: backward left if not just_copy: i = (n - 1) // 2 while i >= 0 and k[i] == '9': i -= 1 if i >= 0: palin[i] = str(int(k[i]) + 1) - for j in range(i + 1, mid): ? ^^^ + for j in range(i + 1, (n + 1) // 2): ? ^^^^^^^^^^^^ palin[j] = '0' - for j in range(mid, n): ? ^^^ + for j in range((n + 1) // 2, n): ? ^^^^^^^^^^^^ mirrored = n - 1 - j palin[j] = palin[mirrored] else: # case 3: "99...9" -> "100..01" palin = ['0'] * (n + 1) palin[0] = palin[-1] = '1' return ''.join(palin) if __name__ == '__main__': t = int(input()) for _ in range(t): k = input() print(next_palindrome(k))
aa014a472a39c12cf3141dd337ecc2ed1ea2cd55
django_summernote/test_settings.py
django_summernote/test_settings.py
import django DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } __MIDDLEWARE__ = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) if django.VERSION < (1, 10): MIDDLEWARE_CLASSES = __MIDDLEWARE__ else: MIDDLEWARE = __MIDDLEWARE__ STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN'
DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } MIDDLEWARE = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN'
Drop old django version support (1.10)
Drop old django version support (1.10)
Python
mit
summernote/django-summernote,summernote/django-summernote,summernote/django-summernote
- import django - DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } - __MIDDLEWARE__ = ( + MIDDLEWARE = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) - - if django.VERSION < (1, 10): - MIDDLEWARE_CLASSES = __MIDDLEWARE__ - else: - MIDDLEWARE = __MIDDLEWARE__ STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN'
Drop old django version support (1.10)
## Code Before: import django DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } __MIDDLEWARE__ = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) if django.VERSION < (1, 10): MIDDLEWARE_CLASSES = __MIDDLEWARE__ else: MIDDLEWARE = __MIDDLEWARE__ STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN' ## Instruction: Drop old django version support (1.10) ## Code After: DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } MIDDLEWARE = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN'
- import django - DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'django_summernote.db', } } - __MIDDLEWARE__ = ( ? -- -- + MIDDLEWARE = ( 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.locale.LocaleMiddleware', ) - - if django.VERSION < (1, 10): - MIDDLEWARE_CLASSES = __MIDDLEWARE__ - else: - MIDDLEWARE = __MIDDLEWARE__ STATIC_URL = '/' MEDIA_URL = '/media/' MEDIA_ROOT = 'test_media' SECRET_KEY = 'django_summernote' ROOT_URLCONF = 'django_summernote.urls' INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.staticfiles', 'django_summernote', ] TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ] }, }, ] if django.VERSION >= (3, 0): X_FRAME_OPTIONS = 'SAMEORIGIN'
8237291e194aa900857fe382d0b8cefb7806c331
ocradmin/ocrmodels/models.py
ocradmin/ocrmodels/models.py
from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) def __unicode__(self): """ String representation. """ return self.name
from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) def __unicode__(self): """ String representation. """ return "<%s: %s>" % (self.__class__.__name__, self.name)
Improve unicode method. Whitespace cleanup
Improve unicode method. Whitespace cleanup
Python
apache-2.0
vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium
from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) - def __unicode__(self): """ String representation. """ - return self.name + return "<%s: %s>" % (self.__class__.__name__, self.name) - +
Improve unicode method. Whitespace cleanup
## Code Before: from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) def __unicode__(self): """ String representation. """ return self.name ## Instruction: Improve unicode method. Whitespace cleanup ## Code After: from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) def __unicode__(self): """ String representation. """ return "<%s: %s>" % (self.__class__.__name__, self.name)
from django.db import models from django.contrib.auth.models import User from tagging.fields import TagField import tagging # OCR model, erm, model class OcrModel(models.Model): """ OCR model objects. """ user = models.ForeignKey(User) derived_from = models.ForeignKey("self", null=True, blank=True) tags = TagField() name = models.CharField(max_length=100, unique=True) description = models.TextField(null=True, blank=True) created_on = models.DateField(auto_now_add=True) updated_on = models.DateField(null=True, blank=True) public = models.BooleanField(default=True) file = models.FileField(upload_to="models") type = models.CharField(max_length=20, choices=[("char", "Character"), ("lang", "Language")]) app = models.CharField(max_length=20, choices=[("ocropus", "Ocropus"), ("tesseract", "Tesseract")]) - def __unicode__(self): """ String representation. """ - return self.name + return "<%s: %s>" % (self.__class__.__name__, self.name) - +
68a7fd8a444a8c568d716db11849f58ad7a9dee5
django_pesapal/views.py
django_pesapal/views.py
from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs)
from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
Add support for further processing of the payment while maintaining compatibility
Add support for further processing of the payment while maintaining compatibility
Python
bsd-3-clause
odero/django-pesapal,odero/django-pesapal
- from django.core.urlresolvers import reverse_lazy + from django.core.urlresolvers import reverse_lazy, reverse - from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView - from django.db.models.loading import get_model + from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False - url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) + url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' - transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') + self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') - merchant_reference = request.GET.get('pesapal_merchant_reference', '') + self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') - if transaction_id and merchant_reference: + if self.transaction_id and self.merchant_reference: - transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) + transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, + pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) + def get_redirect_url(self, *args, **kwargs): + + try: + url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) + except NoReverseMatch: + url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, + kwargs={'merchant_reference': self.merchant_reference}) + return url +
Add support for further processing of the payment while maintaining compatibility
## Code Before: from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) ## Instruction: Add support for further processing of the payment while maintaining compatibility ## Code After: from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
- from django.core.urlresolvers import reverse_lazy + from django.core.urlresolvers import reverse_lazy, reverse ? +++++++++ - from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView - from django.db.models.loading import get_model + from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False - url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) + url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' - transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') + self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') ? +++++ - merchant_reference = request.GET.get('pesapal_merchant_reference', '') + self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') ? +++++ - if transaction_id and merchant_reference: + if self.transaction_id and self.merchant_reference: ? +++++ +++++ - transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) ? ------------------------------------ + transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, ? +++++ + pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) + + def get_redirect_url(self, *args, **kwargs): + + try: + url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) + except NoReverseMatch: + url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, + kwargs={'merchant_reference': self.merchant_reference}) + return url
0a69133e44810dd0469555f62ec49eba120e6ecc
apps/storybase/utils.py
apps/storybase/utils.py
"""Shared utility functions""" from django.template.defaultfilters import slugify as django_slugify def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-')
"""Shared utility functions""" from django.conf import settings from django.template.defaultfilters import slugify as django_slugify from django.utils.translation import ugettext as _ def get_language_name(language_code): """Convert a language code into its full (localized) name""" languages = dict(settings.LANGUAGES) return _(languages[language_code]) def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-')
Add utility function to convert a language code to a its full name
Add utility function to convert a language code to a its full name
Python
mit
denverfoundation/storybase,denverfoundation/storybase,denverfoundation/storybase,denverfoundation/storybase
"""Shared utility functions""" + from django.conf import settings from django.template.defaultfilters import slugify as django_slugify + from django.utils.translation import ugettext as _ + + def get_language_name(language_code): + """Convert a language code into its full (localized) name""" + languages = dict(settings.LANGUAGES) + return _(languages[language_code]) + def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-')
Add utility function to convert a language code to a its full name
## Code Before: """Shared utility functions""" from django.template.defaultfilters import slugify as django_slugify def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-') ## Instruction: Add utility function to convert a language code to a its full name ## Code After: """Shared utility functions""" from django.conf import settings from django.template.defaultfilters import slugify as django_slugify from django.utils.translation import ugettext as _ def get_language_name(language_code): """Convert a language code into its full (localized) name""" languages = dict(settings.LANGUAGES) return _(languages[language_code]) def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-')
"""Shared utility functions""" + from django.conf import settings from django.template.defaultfilters import slugify as django_slugify + from django.utils.translation import ugettext as _ + + def get_language_name(language_code): + """Convert a language code into its full (localized) name""" + languages = dict(settings.LANGUAGES) + return _(languages[language_code]) + def slugify(value): """ Normalizes string, converts to lowercase, removes non-alpha characters, converts spaces to hyphens, and truncates to 50 characters. """ slug = django_slugify(value) slug = slug[:50] return slug.rstrip('-')
a58c3cbfa2c0147525e1afb355e355a9edeb22f8
discussion/admin.py
discussion/admin.py
from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): exclude = ('user',) extra = 1 model = Comment class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment)
from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): extra = 1 model = Comment raw_id_fields = ('user',) class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment)
Add user back onto the comment inline for posts
Add user back onto the comment inline for posts
Python
bsd-2-clause
lehins/lehins-discussion,lehins/lehins-discussion,incuna/django-discussion,incuna/django-discussion,lehins/lehins-discussion
from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): - exclude = ('user',) extra = 1 model = Comment + raw_id_fields = ('user',) class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment)
Add user back onto the comment inline for posts
## Code Before: from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): exclude = ('user',) extra = 1 model = Comment class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment) ## Instruction: Add user back onto the comment inline for posts ## Code After: from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): extra = 1 model = Comment raw_id_fields = ('user',) class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment)
from django.contrib import admin from discussion.models import Comment, Discussion, Post class CommentInline(admin.TabularInline): - exclude = ('user',) extra = 1 model = Comment + raw_id_fields = ('user',) class PostAdmin(admin.ModelAdmin): inlines = (CommentInline,) list_filter = ('discussion',) class DiscussionAdmin(admin.ModelAdmin): prepopulated_fields = { 'slug': ('name',) } admin.site.register(Discussion, DiscussionAdmin) admin.site.register(Post, PostAdmin) admin.site.register(Comment)
1e5e2a236277dc9ba11f9fe4aff3279f692da3f7
ploy/tests/conftest.py
ploy/tests/conftest.py
from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock
from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) def content(self): with open(self.path) as f: return f.read() @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock
Add convenience function to read tempdir files.
Add convenience function to read tempdir files.
Python
bsd-3-clause
fschulze/ploy,ployground/ploy
from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) + def content(self): + with open(self.path) as f: + return f.read() + @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock
Add convenience function to read tempdir files.
## Code Before: from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock ## Instruction: Add convenience function to read tempdir files. ## Code After: from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) def content(self): with open(self.path) as f: return f.read() @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock
from mock import patch import pytest import os import shutil import tempfile class Directory: def __init__(self, directory): self.directory = directory def __getitem__(self, name): path = os.path.join(self.directory, name) assert not os.path.relpath(path, self.directory).startswith('..') return File(path) class File: def __init__(self, path): self.directory = os.path.dirname(path) self.path = path def fill(self, content): if not os.path.exists(self.directory): os.makedirs(self.directory) with open(self.path, 'w') as f: if isinstance(content, (list, tuple)): content = '\n'.join(content) f.write(content) + def content(self): + with open(self.path) as f: + return f.read() + @pytest.yield_fixture def tempdir(): """ Returns an object for easy use of a temporary directory which is cleaned up afterwards. Use tempdir[filepath] to access files. Use .fill(lines) on the returned object to write content to the file. """ directory = tempfile.mkdtemp() yield Directory(directory) shutil.rmtree(directory) @pytest.yield_fixture def ployconf(tempdir): """ Returns a Configfile object which manages ploy.conf. """ yield tempdir['etc/ploy.conf'] @pytest.yield_fixture def os_execvp_mock(): with patch("os.execvp") as os_execvp_mock: yield os_execvp_mock
7a374b19cf89421a73ea55fdbcd1b16b52327568
dm_control/composer/initializer.py
dm_control/composer/initializer.py
"""Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod def initialize_episode(self, physics, random_state): raise NotImplementedError
"""Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod def __call__(self, physics, random_state): raise NotImplementedError
Rename `initialize_episode` --> `__call__` in `composer.Initializer`
Rename `initialize_episode` --> `__call__` in `composer.Initializer` PiperOrigin-RevId: 234775654
Python
apache-2.0
deepmind/dm_control
"""Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod - def initialize_episode(self, physics, random_state): + def __call__(self, physics, random_state): raise NotImplementedError
Rename `initialize_episode` --> `__call__` in `composer.Initializer`
## Code Before: """Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod def initialize_episode(self, physics, random_state): raise NotImplementedError ## Instruction: Rename `initialize_episode` --> `__call__` in `composer.Initializer` ## Code After: """Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod def __call__(self, physics, random_state): raise NotImplementedError
"""Module defining the abstract initializer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import six @six.add_metaclass(abc.ABCMeta) class Initializer(object): """The abstract base class for an initializer.""" @abc.abstractmethod - def initialize_episode(self, physics, random_state): ? ^^^^^ ^^^ ^^^^^^^ + def __call__(self, physics, random_state): ? ^^^ ^ ^ raise NotImplementedError
2c357a54e30eecb1d7b717be3ed774dcfecc2814
src/stratis_cli/_actions/_stratis.py
src/stratis_cli/_actions/_stratis.py
from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod def list_stratisd_redundancy(namespace): """ List the stratisd redundancy designations. """ # pylint: disable=unused-argument for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod def list_stratisd_version(namespace): """ List the stratisd version. """ # pylint: disable=unused-argument print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod def list_stratisd_redundancy(_namespace): """ List the stratisd redundancy designations. """ for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod def list_stratisd_version(_namespace): """ List the stratisd version. """ print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
Use '_' prefix instead of disabling pylint unused-argument lint
Use '_' prefix instead of disabling pylint unused-argument lint It is more precise to mark the unused parameters this way. Signed-off-by: mulhern <7b51bcf507bcd7afb72bf8663752c0ddbeb517f6@redhat.com>
Python
apache-2.0
stratis-storage/stratis-cli,stratis-storage/stratis-cli
from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod - def list_stratisd_redundancy(namespace): + def list_stratisd_redundancy(_namespace): """ List the stratisd redundancy designations. """ - # pylint: disable=unused-argument for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod - def list_stratisd_version(namespace): + def list_stratisd_version(_namespace): """ List the stratisd version. """ - # pylint: disable=unused-argument print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
Use '_' prefix instead of disabling pylint unused-argument lint
## Code Before: from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod def list_stratisd_redundancy(namespace): """ List the stratisd redundancy designations. """ # pylint: disable=unused-argument for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod def list_stratisd_version(namespace): """ List the stratisd version. """ # pylint: disable=unused-argument print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT))) ## Instruction: Use '_' prefix instead of disabling pylint unused-argument lint ## Code After: from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod def list_stratisd_redundancy(_namespace): """ List the stratisd redundancy designations. """ for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod def list_stratisd_version(_namespace): """ List the stratisd version. """ print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
from .._stratisd_constants import RedundancyCodes from ._connection import get_object from ._constants import TOP_OBJECT from ._data import Manager class StratisActions(): """ Stratis actions. """ @staticmethod - def list_stratisd_redundancy(namespace): + def list_stratisd_redundancy(_namespace): ? + """ List the stratisd redundancy designations. """ - # pylint: disable=unused-argument for code in RedundancyCodes: print("%s: %d" % (code.name, code.value)) @staticmethod - def list_stratisd_version(namespace): + def list_stratisd_version(_namespace): ? + """ List the stratisd version. """ - # pylint: disable=unused-argument print("%s" % Manager.Properties.Version.Get(get_object(TOP_OBJECT)))
d4d73fe7d5e83c65d9abbf59ea14ed60eb23a83f
poem_reader.py
poem_reader.py
import argparse from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") args = argparser.parse_args() data_root = args.dataroot with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f: tree = etree.parse(f) root = tree.getroot() print(root.tag)
import argparse import glob from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") args = argparser.parse_args() data_root = args.dataroot def read_xml_directory(path): """ Read XML files from path, parse them, and return them as list """ files = glob.glob(path + "*.xml") xmls = [] for xmlfile in files: with open(xmlfile, 'r') as f: xmls.append(etree.parse(f)) return xmls def find_by_block_id(xmls, block_id): """ Find an element by block_id from a list of lxml trees """ block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id)) for xml in xmls: elements = block_xpath(xml) if elements: return elements[0] some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/' xmls = read_xml_directory(some_dir) print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
Read XML files from a directory and find textblock by id
Read XML files from a directory and find textblock by id
Python
mit
dhh17/categories_norms_genres,dhh17/categories_norms_genres,dhh17/categories_norms_genres
import argparse + import glob from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') + argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") + args = argparser.parse_args() - argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") - - args = argparser.parse_args() data_root = args.dataroot - with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f: - tree = etree.parse(f) - root = tree.getroot() - print(root.tag) + def read_xml_directory(path): + """ + Read XML files from path, parse them, and return them as list + """ + files = glob.glob(path + "*.xml") + xmls = [] + for xmlfile in files: + with open(xmlfile, 'r') as f: + xmls.append(etree.parse(f)) + + return xmls + + + def find_by_block_id(xmls, block_id): + """ + Find an element by block_id from a list of lxml trees + """ + block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id)) + for xml in xmls: + elements = block_xpath(xml) + if elements: + return elements[0] + + + some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/' + xmls = read_xml_directory(some_dir) + + print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001'))) + + +
Read XML files from a directory and find textblock by id
## Code Before: import argparse from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") args = argparser.parse_args() data_root = args.dataroot with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f: tree = etree.parse(f) root = tree.getroot() print(root.tag) ## Instruction: Read XML files from a directory and find textblock by id ## Code After: import argparse import glob from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") args = argparser.parse_args() data_root = args.dataroot def read_xml_directory(path): """ Read XML files from path, parse them, and return them as list """ files = glob.glob(path + "*.xml") xmls = [] for xmlfile in files: with open(xmlfile, 'r') as f: xmls.append(etree.parse(f)) return xmls def find_by_block_id(xmls, block_id): """ Find an element by block_id from a list of lxml trees """ block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id)) for xml in xmls: elements = block_xpath(xml) if elements: return elements[0] some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/' xmls = read_xml_directory(some_dir) print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001')))
import argparse + import glob from lxml import etree argparser = argparse.ArgumentParser(description="Newspaper XML parser", fromfile_prefix_chars='@') + argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") + args = argparser.parse_args() - argparser.add_argument("dataroot", help="Path to DHH 17 newspapers directory") - - args = argparser.parse_args() data_root = args.dataroot - with open(data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/1457-4616_1854-08-01_31_001.xml', 'r') as f: - tree = etree.parse(f) - root = tree.getroot() - print(root.tag) + def read_xml_directory(path): + """ + Read XML files from path, parse them, and return them as list + """ + files = glob.glob(path + "*.xml") + + xmls = [] + for xmlfile in files: + with open(xmlfile, 'r') as f: + xmls.append(etree.parse(f)) + + return xmls + + + def find_by_block_id(xmls, block_id): + """ + Find an element by block_id from a list of lxml trees + """ + block_xpath = etree.XPath("//*[@ID='{id}']".format(id=block_id)) + for xml in xmls: + elements = block_xpath(xml) + if elements: + return elements[0] + + + some_dir = data_root + 'newspapers/fin/1854/1457-4616/1457-4616_1854-08-01_31/alto/' + xmls = read_xml_directory(some_dir) + + print(etree.tostring(find_by_block_id(xmls, 'P2_TB00001'))) + +
b583c5fb00d1ebfa0458a6233be85d8b56173abf
python/printbag.py
python/printbag.py
import sys import logging import numpy as np # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag from antlia.dtype import LIDAR_CONVERTED_DTYPE def print_bag(bag, topics=None): if topics is None: #topics = ['/tf', '/scan'] topics = ['/scan', '/flagbutton_pressed'] for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: print(('Usage: {} <rosbag> \n\n' 'Print contents of rosbag file.' ).format(__file__)) sys.exit(1) outfile = None filename = sys.argv[1] with rosbag.Bag(filename) as bag: print_bag(bag) sys.exit()
import sys import logging # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag def print_bag(bag, topics=None): for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: print(('Usage: {} [topics] <rosbag> \n\n' 'topics:\tcomma-separated list of topics\n\n' 'Print contents of rosbag file. If topics is not provided, \n' 'all topics are printed\n' ).format(__file__)) sys.exit(1) topics = None if len(sys.argv) == 3: topics = [t.strip() for t in sys.argv[1].split(',')] filename = sys.argv[2] else: filename = sys.argv[1] with rosbag.Bag(filename) as bag: print_bag(bag, topics) sys.exit()
Add argument to specify bag topics
Add argument to specify bag topics
Python
bsd-2-clause
oliverlee/antlia
- import sys import logging - - import numpy as np # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag - from antlia.dtype import LIDAR_CONVERTED_DTYPE def print_bag(bag, topics=None): - if topics is None: - #topics = ['/tf', '/scan'] - topics = ['/scan', '/flagbutton_pressed'] - for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: - print(('Usage: {} <rosbag> \n\n' + print(('Usage: {} [topics] <rosbag> \n\n' + 'topics:\tcomma-separated list of topics\n\n' - 'Print contents of rosbag file.' + 'Print contents of rosbag file. If topics is not provided, \n' + 'all topics are printed\n' ).format(__file__)) sys.exit(1) - outfile = None + topics = None + if len(sys.argv) == 3: + topics = [t.strip() for t in sys.argv[1].split(',')] + filename = sys.argv[2] + else: - filename = sys.argv[1] + filename = sys.argv[1] with rosbag.Bag(filename) as bag: - print_bag(bag) + print_bag(bag, topics) sys.exit()
Add argument to specify bag topics
## Code Before: import sys import logging import numpy as np # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag from antlia.dtype import LIDAR_CONVERTED_DTYPE def print_bag(bag, topics=None): if topics is None: #topics = ['/tf', '/scan'] topics = ['/scan', '/flagbutton_pressed'] for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: print(('Usage: {} <rosbag> \n\n' 'Print contents of rosbag file.' ).format(__file__)) sys.exit(1) outfile = None filename = sys.argv[1] with rosbag.Bag(filename) as bag: print_bag(bag) sys.exit() ## Instruction: Add argument to specify bag topics ## Code After: import sys import logging # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag def print_bag(bag, topics=None): for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: print(('Usage: {} [topics] <rosbag> \n\n' 'topics:\tcomma-separated list of topics\n\n' 'Print contents of rosbag file. If topics is not provided, \n' 'all topics are printed\n' ).format(__file__)) sys.exit(1) topics = None if len(sys.argv) == 3: topics = [t.strip() for t in sys.argv[1].split(',')] filename = sys.argv[2] else: filename = sys.argv[1] with rosbag.Bag(filename) as bag: print_bag(bag, topics) sys.exit()
- import sys import logging - - import numpy as np # suppress logging warnings due to rospy logging.basicConfig(filename='/dev/null') import rosbag - from antlia.dtype import LIDAR_CONVERTED_DTYPE def print_bag(bag, topics=None): - if topics is None: - #topics = ['/tf', '/scan'] - topics = ['/scan', '/flagbutton_pressed'] - for message in bag.read_messages(topics=topics): print(message) if __name__ == '__main__': if len(sys.argv) < 2: - print(('Usage: {} <rosbag> \n\n' + print(('Usage: {} [topics] <rosbag> \n\n' ? +++++++++ + 'topics:\tcomma-separated list of topics\n\n' - 'Print contents of rosbag file.' + 'Print contents of rosbag file. If topics is not provided, \n' ? ++++++++++++++++++++++++++++++ + 'all topics are printed\n' ).format(__file__)) sys.exit(1) - outfile = None + topics = None + if len(sys.argv) == 3: + topics = [t.strip() for t in sys.argv[1].split(',')] + filename = sys.argv[2] + else: - filename = sys.argv[1] + filename = sys.argv[1] ? ++++ with rosbag.Bag(filename) as bag: - print_bag(bag) + print_bag(bag, topics) ? ++++++++ sys.exit()
edf38ad11631ad5e793eb9ac95dbc865595d517b
glue_vispy_viewers/common/layer_state.py
glue_vispy_viewers/common/layer_state.py
from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State from glue.core.message import LayerArtistUpdatedMessage __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() self.add_global_callback(self._notify_layer_update) def _notify_layer_update(self, **kwargs): message = LayerArtistUpdatedMessage(self) if self.layer is not None and self.layer.hub is not None: self.layer.hub.broadcast(message) def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
Make sure layer artist icon updates when changing the color mode or colormaps
Make sure layer artist icon updates when changing the color mode or colormaps
Python
bsd-2-clause
glue-viz/glue-vispy-viewers,PennyQ/astro-vispy,astrofrog/glue-3d-viewer,glue-viz/glue-3d-viewer,astrofrog/glue-vispy-viewers
from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State + from glue.core.message import LayerArtistUpdatedMessage __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() + self.add_global_callback(self._notify_layer_update) + + def _notify_layer_update(self, **kwargs): + message = LayerArtistUpdatedMessage(self) + if self.layer is not None and self.layer.hub is not None: + self.layer.hub.broadcast(message) + def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
Make sure layer artist icon updates when changing the color mode or colormaps
## Code Before: from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha') ## Instruction: Make sure layer artist icon updates when changing the color mode or colormaps ## Code After: from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State from glue.core.message import LayerArtistUpdatedMessage __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() self.add_global_callback(self._notify_layer_update) def _notify_layer_update(self, **kwargs): message = LayerArtistUpdatedMessage(self) if self.layer is not None and self.layer.hub is not None: self.layer.hub.broadcast(message) def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
from __future__ import absolute_import, division, print_function from glue.external.echo import CallbackProperty, keep_in_sync from glue.core.state_objects import State + from glue.core.message import LayerArtistUpdatedMessage __all__ = ['VispyLayerState'] class VispyLayerState(State): """ A base state object for all Vispy layers """ layer = CallbackProperty() visible = CallbackProperty(True) zorder = CallbackProperty(0) color = CallbackProperty() alpha = CallbackProperty() def __init__(self, **kwargs): super(VispyLayerState, self).__init__(**kwargs) self._sync_color = None self._sync_alpha = None self.add_callback('layer', self._layer_changed) self._layer_changed() + self.add_global_callback(self._notify_layer_update) + + def _notify_layer_update(self, **kwargs): + message = LayerArtistUpdatedMessage(self) + if self.layer is not None and self.layer.hub is not None: + self.layer.hub.broadcast(message) + def _layer_changed(self): if self._sync_color is not None: self._sync_color.stop_syncing() if self._sync_alpha is not None: self._sync_alpha.stop_syncing() if self.layer is not None: self.color = self.layer.style.color self.alpha = self.layer.style.alpha self._sync_color = keep_in_sync(self, 'color', self.layer.style, 'color') self._sync_alpha = keep_in_sync(self, 'alpha', self.layer.style, 'alpha')
e836f3c558085aa0a1275546ac45b8146254ee6b
test/default.py
test/default.py
from mock import MagicMock import pbclient class TestDefault(object): """Test class for pbs.helpers.""" error = {"action": "GET", "exception_cls": "NotFound", "exception_msg": "(NotFound)", "status": "failed", "status_code": 404, "target": "/api/app"} config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', 'description': 'description', 'short_name': 'short_name'} def tearDown(self): self.error['status'] = 'failed'
"""Test module for pbs client.""" from mock import MagicMock import pbclient class TestDefault(object): """Test class for pbs.helpers.""" config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', 'description': 'description', 'short_name': 'short_name'} def tearDown(self): """Tear down method.""" self.error['status'] = 'failed' @property def error(self, action='GET', exception_cls='NotFound', exception_msg='(NotFound)', status='failed', status_code=404, target='/api/app'): """Error property.""" return {'action': action, 'exception_cls': exception_cls, 'exception_msg': exception_msg, 'status': status, 'status_code': status_code, 'target': target}
Refactor error as a property.
Refactor error as a property.
Python
agpl-3.0
PyBossa/pbs,PyBossa/pbs,PyBossa/pbs
+ """Test module for pbs client.""" from mock import MagicMock import pbclient + + class TestDefault(object): """Test class for pbs.helpers.""" - - error = {"action": "GET", - "exception_cls": "NotFound", - "exception_msg": "(NotFound)", - "status": "failed", - "status_code": 404, - "target": "/api/app"} config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', - 'description': 'description', + 'description': 'description', - 'short_name': 'short_name'} + 'short_name': 'short_name'} def tearDown(self): + """Tear down method.""" self.error['status'] = 'failed' + @property + def error(self, action='GET', + exception_cls='NotFound', + exception_msg='(NotFound)', + status='failed', + status_code=404, + target='/api/app'): + """Error property.""" + return {'action': action, + 'exception_cls': exception_cls, + 'exception_msg': exception_msg, + 'status': status, + 'status_code': status_code, + 'target': target} +
Refactor error as a property.
## Code Before: from mock import MagicMock import pbclient class TestDefault(object): """Test class for pbs.helpers.""" error = {"action": "GET", "exception_cls": "NotFound", "exception_msg": "(NotFound)", "status": "failed", "status_code": 404, "target": "/api/app"} config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', 'description': 'description', 'short_name': 'short_name'} def tearDown(self): self.error['status'] = 'failed' ## Instruction: Refactor error as a property. ## Code After: """Test module for pbs client.""" from mock import MagicMock import pbclient class TestDefault(object): """Test class for pbs.helpers.""" config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', 'description': 'description', 'short_name': 'short_name'} def tearDown(self): """Tear down method.""" self.error['status'] = 'failed' @property def error(self, action='GET', exception_cls='NotFound', exception_msg='(NotFound)', status='failed', status_code=404, target='/api/app'): """Error property.""" return {'action': action, 'exception_cls': exception_cls, 'exception_msg': exception_msg, 'status': status, 'status_code': status_code, 'target': target}
+ """Test module for pbs client.""" from mock import MagicMock import pbclient + + class TestDefault(object): """Test class for pbs.helpers.""" - - error = {"action": "GET", - "exception_cls": "NotFound", - "exception_msg": "(NotFound)", - "status": "failed", - "status_code": 404, - "target": "/api/app"} config = MagicMock() config.server = 'http://server' config.api_key = 'apikey' config.pbclient = pbclient config.project = {'name': 'name', - 'description': 'description', + 'description': 'description', ? +++++ - 'short_name': 'short_name'} + 'short_name': 'short_name'} ? +++++ def tearDown(self): + """Tear down method.""" self.error['status'] = 'failed' + + @property + def error(self, action='GET', + exception_cls='NotFound', + exception_msg='(NotFound)', + status='failed', + status_code=404, + target='/api/app'): + """Error property.""" + return {'action': action, + 'exception_cls': exception_cls, + 'exception_msg': exception_msg, + 'status': status, + 'status_code': status_code, + 'target': target}
fbc5e2d52549452c2adbe58644358cf3c4eeb526
testsuite/test_util.py
testsuite/test_util.py
import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*'])
import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) self.assertEquals(pep8.normalize_paths([]), []) self.assertEquals(pep8.normalize_paths(None), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*'])
Add a few more cases of "not value"
Add a few more cases of "not value"
Python
mit
ojengwa/pep8,pedros/pep8,asandyz/pep8,jayvdb/pep8,doismellburning/pep8,pandeesh/pep8,jayvdb/pep8,PyCQA/pep8,ABaldwinHunter/pep8,codeclimate/pep8,ABaldwinHunter/pep8-clone-classic,zevnux/pep8,MeteorAdminz/pep8
import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) + self.assertEquals(pep8.normalize_paths([]), []) + self.assertEquals(pep8.normalize_paths(None), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) + self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*'])
Add a few more cases of "not value"
## Code Before: import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*']) ## Instruction: Add a few more cases of "not value" ## Code After: import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) self.assertEquals(pep8.normalize_paths([]), []) self.assertEquals(pep8.normalize_paths(None), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*'])
import os import unittest import pep8 class UtilTestCase(unittest.TestCase): def test_normalize_paths(self): cwd = os.getcwd() self.assertEquals(pep8.normalize_paths(''), []) + self.assertEquals(pep8.normalize_paths([]), []) + self.assertEquals(pep8.normalize_paths(None), []) self.assertEquals(pep8.normalize_paths(['foo']), ['foo']) self.assertEquals(pep8.normalize_paths('foo'), ['foo']) self.assertEquals(pep8.normalize_paths('foo,bar'), ['foo', 'bar']) + self.assertEquals(pep8.normalize_paths('foo, bar '), ['foo', 'bar']) self.assertEquals(pep8.normalize_paths('/foo/bar,baz/../bat'), ['/foo/bar', cwd + '/bat']) self.assertEquals(pep8.normalize_paths(".pyc,\n build/*"), ['.pyc', cwd + '/build/*'])