code stringlengths 114 1.05M | path stringlengths 3 312 | quality_prob float64 0.5 0.99 | learning_prob float64 0.2 1 | filename stringlengths 3 168 | kind stringclasses 1
value |
|---|---|---|---|---|---|
import random
import re
from question_builder.bp.exceptions.question_creator_exception import (
QuestionCreatorException,
)
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/future_continuous/future_continuous_negative_choice_question_creator.py | 0.514156 | 0.299707 | future_continuous_negative_choice_question_creator.py | pypi |
import random
import re
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
PATTERN_KEY = "pattern"
SUBJECT_KEY = "subject"
VB = "VB"
BEEN = "been"
BEEING = "beeing"
BE = "be"
N_BA... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/passive_voice/passive_voice_choice_question_creator.py | 0.433022 | 0.326084 | passive_voice_choice_question_creator.py | pypi |
import random
import editdistance
from question_builder.bp.exceptions.lexemas_exceptions import NoLexemasFound
from question_builder.bp.exceptions.pos_exceptions import WordNotVerb
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Que... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/simple_present/simple_tense_choice_question_creator.py | 0.428233 | 0.370567 | simple_tense_choice_question_creator.py | pypi |
import random
from question_builder.bp.dictionary_factory import word2englishdefinitions
from question_builder.bp.exceptions.dictionary_exceptions import (
WordNotInEnglishDefinitionsDictionary,
)
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/simple_present/simple_present_negation_choice_question_creator.py | 0.594669 | 0.309115 | simple_present_negation_choice_question_creator.py | pypi |
import random
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
SUBJECT_KEY = "subject"
CONJUGATED_VERBTOBE_KEY = "conjugated_auxiliaryverb"
TARGET_VERB_CONJUGATIONS_KEY = "targe... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/past_continuous/past_continuous_choice_question_creator.py | 0.569015 | 0.291252 | past_continuous_choice_question_creator.py | pypi |
import random
from question_builder.bp.exceptions.question_creator_exception import (
QuestionCreatorException,
)
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
SUBJECT_KE... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/past_continuous/past_continuous_negation_choice_question_creator.py | 0.591841 | 0.292179 | past_continuous_negation_choice_question_creator.py | pypi |
import random
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
SUBJECT_KEY = "subject"
TARGET_VERB_CONJUGATIONS_KEY = "lemma_conjugations"
PRESENT_CONTINUOUS_KEY = "VBG"
HAD = "... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/past_perfect/past_perfect_choice_question_creator.py | 0.546012 | 0.275615 | past_perfect_choice_question_creator.py | pypi |
import random
from question_builder.bp.question_creators.question_creator import QuestionCreator
from question_builder.bp.questions.question import Question
from question_builder.data import DataQuestion
SUBJECT_KEY = "subject"
NEGATION_ADVERB_KEY = "negation"
HAD = "had"
HAVE = "have"
NOT = "not"
NO = "no"
DIDNT = "... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/bp/question_creators/patterns/verbs/past_perfect/past_perfect_negation_choice_question_creator.py | 0.556159 | 0.327776 | past_perfect_negation_choice_question_creator.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'FDV'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 4}]-(subject:WORD)
MATCH (content)... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/going_to_future/going_to_future_affirmative_queries.py | 0.437583 | 0.160891 | going_to_future_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'GTFNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 5}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/going_to_future/going_to_future_negative_queries.py | 0.472683 | 0.16975 | going_to_future_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PPNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (content... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_perfect/present_perfect_negative_queries.py | 0.472927 | 0.177454 | present_perfect_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PPC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)-... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_perfect/present_perfect_affirmative_queries.py | 0.436382 | 0.166337 | present_perfect_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:\\"PCC\\"})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 4}]-(subject:WORD)
MATCH (cont... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_perfect_continuous/future_perfect_continuous_affirmative_queries.py | 0.446495 | 0.159283 | future_perfect_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:\\"PCNC\\"})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 5}]-(subject:WORD)
MATCH (con... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_perfect_continuous/future_perfect_continuous_negative_queries.py | 0.485356 | 0.163445 | future_perfect_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'SFC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)-... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/simple_future/simple_future_affirmative_queries.py | 0.441191 | 0.168412 | simple_future_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'SFNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (content... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/simple_future/simple_future_negative_queries.py | 0.47025 | 0.174516 | simple_future_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PCNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (content... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_continuous/present_continuous_negative_queries.py | 0.469763 | 0.172869 | present_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PCIPT'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_continuous/present_continuous_interrogative_queries.py | 0.46223 | 0.168446 | present_continuous_interrogative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PCC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)-... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_continuous/present_continuous_affirmative_queries.py | 0.438304 | 0.167083 | present_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'FPC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 3}]-(subject:WORD)
MATCH (content)-... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_perfect/future_perfect_affirmative_queries.py | 0.441673 | 0.161419 | future_perfect_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'FPNC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 4}]-(subject:WORD)
MATCH (content)... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_perfect/future_perfect_negative_queries.py | 0.452052 | 0.160003 | future_perfect_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PRPCC'})
MATCH (c)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (c)-[:HAS_WOR... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_perfect_continuous/present_perfect_continuous_affirmative_queries.py | 0.438304 | 0.166913 | present_perfect_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PRPCNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 4}]-(subject:WORD)
MATCH (conte... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/present_perfect_continuous/present_perfect_continuous_negative_queries.py | 0.471953 | 0.168446 | present_perfect_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:\\"PPCC\\"})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 3}]-(subject:WORD)
MATCH (cont... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_perfect_continuous/past_perfect_continuous_affirmative_queries.py | 0.448909 | 0.156846 | past_perfect_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PPCNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 4}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_perfect_continuous/past_perfect_continuous_negative_queries.py | 0.474875 | 0.170508 | past_perfect_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'SPANC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/simple_past/simple_past_negative_queries.py | 0.461988 | 0.182863 | simple_past_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'FCC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (content)... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_continuous/future_continuous_affirmative_queries.py | 0.430147 | 0.169372 | future_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'FCNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 4}]-(subject:WORD)
MATCH (content... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/future_continuous/future_continuous_negative_queries.py | 0.47926 | 0.170301 | future_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PVC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)-... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/passive_voice/passive_voice_affirmative_queries.py | 0.44071 | 0.166167 | passive_voice_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'SPNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (content... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/simple_present/simple_present_negative_queries.py | 0.461017 | 0.179423 | simple_present_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PACNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_continuous/past_continuous_negative_queries.py | 0.462473 | 0.179746 | past_continuous_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PACC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_continuous/past_continuous_affirmative_queries.py | 0.421552 | 0.171165 | past_continuous_affirmative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PAPNC'})
MATCH (content)-[:HAS_WORD{position : pattern_rel.position - 3}]-(subject:WORD)
MATCH (conten... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_perfect/past_perfect_negative_queries.py | 0.461988 | 0.175715 | past_perfect_negative_queries.py | pypi |
GET_N_DATA_QUESTIONS_PER_WORD = """
UNWIND $lemma_list as lemma
CALL apoc.cypher.run("
MATCH (interest) - [:HAS_CONTENT] -> (content:CONTENT) - [pattern_rel:HAS_PATTERN{lemma:{lemma}}] -> (:PATTERN{code:'PAPC'})
MATCH (content)-[:HAS_WORD{position: pattern_rel.position - 2}]-(subject:WORD)
MATCH (content)... | /questionbuilder-1.1.13.tar.gz/questionbuilder-1.1.13/question_builder/data/question_type_repository/verbs/past_perfect/past_perfect_affirmative_queries.py | 0.438785 | 0.165121 | past_perfect_affirmative_queries.py | pypi |
from .question_creators.antonym_question_creator import AntonymQuestionCreator
from .question_creators.subword_question_creator import SubwordQuestionCreator
from .question_creators.intruder_question_creator import IntruderQuestionCreator
from .question_creators.synonym_question_creator import SynonymQuestionCreator
fr... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_pack.py | 0.523908 | 0.545407 | question_pack.py | pypi |
from question_builder.data import DataQuestion
from .question_creators.question_creator_exception import QuestionCreatorException # noqa: E501
import random
import logging
class QuestionBuilder():
"""
Builds a question given the raw data DataQuestion information.
It uses the responsibility chain design p... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_builder.py | 0.535584 | 0.469155 | question_builder.py | pypi |
from question_builder.data import DataQuestion
from ..questions.question import Question
from .question_creator import QuestionCreator
from ..dictionary_factory import word2partwordtyping, CORRECT
from .question_creator_exception import WordNotInPartWordTypingDictionary
import re
class PartWordTypingQuestionCreator(Qu... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/partword_typing_question_creator.py | 0.41253 | 0.443118 | partword_typing_question_creator.py | pypi |
from question_builder.data import DataQuestion
from . import pos_validators
from . import tag_validators
from ..questions.question import Question
from .question_creator import QuestionCreator
from .question_creator_exception import NoLexemasFound, WordNotVerb
from .question_creator_exception import VerbNotPastTense
im... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/simple_tense_choice_question_creator.py | 0.435421 | 0.381594 | simple_tense_choice_question_creator.py | pypi |
from question_builder.data import DataQuestion
from . import pos_validators
from ..questions.question import Question
from .question_creator import QuestionCreator
from .question_creator_exception import NoLexemasFound, WordNotVerb
import random
import editdistance
class TenseVerbChoiceQuestionCreator(QuestionCreator)... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/tense_verb_choice_question_creator.py | 0.439026 | 0.386937 | tense_verb_choice_question_creator.py | pypi |
from question_builder.data import DataQuestion
from . import pos_validators
from ..questions.question import Question, CORRECT, TEXT, N_BAITS
from .question_creator import QuestionCreator
from ..dictionary_factory import word2presuffix, CORRECT as DICT_CORRECT, BAITS
from .question_creator_exception import WordHasNotVa... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/subword_question_creator.py | 0.414425 | 0.419945 | subword_question_creator.py | pypi |
from question_builder.data import DataQuestion
from ...config import MIN_SENTENCE_SCRABBLE_LENGTH, MAX_SENTENCE_SCRABBLE_LENGTH
from ..questions.question import Question, TEXT, CORRECT
from .question_creator import QuestionCreator
from .question_creator_exception import MaximumLengthExceeded
import re
class SentenceSc... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/sentence_scrabble_question_creator.py | 0.465387 | 0.297151 | sentence_scrabble_question_creator.py | pypi |
from question_builder.data import DataQuestion
from ..questions.question import Question, N_BAITS
from .question_creator import QuestionCreator
from ..dictionary_factory import word2multisynonym, word2soundslike, pos2words
from .question_creator_exception import WordNotInMultisynonymDictionary, WordNotInSoundsLikeDicti... | /questiongeneratorcesar-0.0.9-py3-none-any.whl/question_builder/bp/question_creators/synonym_question_creator.py | 0.425963 | 0.45647 | synonym_question_creator.py | pypi |
# questionnaire

[](https://badge.fury.io/py/questionnaire)
[ for the Questrade API.
2. [Generate](https://login.questrade.com/APIAcce... | /questrade-api-1.0.3.tar.gz/questrade-api-1.0.3/README.md | 0.542379 | 0.863161 | README.md | pypi |
import hashlib
from dataclasses import dataclass
from itertools import count
from pathlib import Path
from typing import Dict, Iterator, List, Optional, Union
import requests
from dacite import from_dict
@dataclass(frozen=True)
class Channel:
name: str
description: str
private: bool
size_limit: Optio... | /quetz_client-0.5.0.tar.gz/quetz_client-0.5.0/src/quetz_client/client.py | 0.842199 | 0.166709 | client.py | pypi |
from __future__ import annotations
import json
import uuid
from datetime import date, datetime
from enum import Enum
from typing import Generic, List, Optional, TypeVar
from pydantic import BaseModel, Field, root_validator, validator
from pydantic.generics import GenericModel
T = TypeVar('T')
class BaseProfile(Ba... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/rest_models.py | 0.916629 | 0.198511 | rest_models.py | pypi |
from pathlib import Path
from typing import BinaryIO, List, Optional
import fastapi
import pluggy
import quetz
from quetz.condainfo import CondaInfo
hookspec = pluggy.HookspecMarker("quetz")
@hookspec
def register_router() -> 'fastapi.APIRouter':
"""add extra endpoints to the url tree.
It should return an... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/hooks.py | 0.907947 | 0.24225 | hooks.py | pypi |
import abc
import base64
import calendar
import contextlib
import datetime
import hashlib
import logging
import os
import os.path as path
import shutil
import tempfile
import warnings
from contextlib import contextmanager
from os import PathLike
from threading import Lock
from typing import IO, List, Tuple, Union
imp... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/pkgstores.py | 0.773601 | 0.152347 | pkgstores.py | pypi |
import uuid
from datetime import datetime, timedelta
from enum import Enum
import sqlalchemy as sa
from quetz.db_models import UUID, Base
class IntervalType(Enum):
hour = "H"
day = "D"
month = "M"
year = "Y"
@property
def timedelta(self):
if self == IntervalType.hour:
re... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/metrics/db_models.py | 0.754553 | 0.289133 | db_models.py | pypi |
from quetz.config import Config
from .oauth2 import OAuthAuthenticator
class GoogleAuthenticator(OAuthAuthenticator):
"""Use Google account to authenticate users with Quetz.
To enable add the following to the configuration file:
.. code::
[google]
client_id = "1111111111-dha39auqzp92110sd... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/authentication/google.py | 0.609873 | 0.17621 | google.py | pypi |
from .oauth2 import OAuthAuthenticator
class GitlabAuthenticator(OAuthAuthenticator):
"""Use Gitlab account to authenticate users with Quetz.
To enable add the following to the configuration file:
.. code::
[gitlab]
client_id = "fde330aef1fbe39991"
client_secret = "03728444a12abff17e9... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/authentication/gitlab.py | 0.707607 | 0.207556 | gitlab.py | pypi |
import json
from .oauth2 import OAuthAuthenticator
class AzureADAuthenticator(OAuthAuthenticator):
"""Use Microsoft Azure Active Directory account to authenticate users with Quetz.
To enable add the following to the configuration file:
.. code::
[azuread]
client_id = "some-client-id-value"... | /quetz-server-0.9.2.tar.gz/quetz-server-0.9.2/quetz/authentication/azuread.py | 0.5564 | 0.164081 | azuread.py | pypi |
def phylip2arlequin(input: str, imap: str, output: str) -> None :
"""Converts a PHYLIP alignement to ALREQUIN format using population mapping
Args:
input : The path to the phylip sequence file - space delimited csv
imap : The path to the imap file mapping sequences ids to population - space del... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/phylip2arlequin.py | 0.522446 | 0.606265 | phylip2arlequin.py | pypi |
import sqlite3, pyvolve, os
from optparse import OptionParser
from . import sequence
def newick_list_to_phylip_simulation(newicks, sequence_size, scale_tree, output, temporary_prefix):
temporaries = []
my_model = pyvolve.Model("nucleotide")
partition = pyvolve.Partition(models = my_model, size = sequence_... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/simulate_phylip_sequences.py | 0.446977 | 0.246806 | simulate_phylip_sequences.py | pypi |
from optparse import OptionParser
import rasterio
import math
import rasterio.mask
from functools import partial
import pyproj
from shapely.ops import transform
from shapely.geometry import Point, Polygon
proj_wgs84 = pyproj.Proj('+proj=longlat +datum=WGS84')
from math import cos, sin, asin, sqrt, radians
def calc_... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/raster/circle_mask.py | 0.797439 | 0.482185 | circle_mask.py | pypi |
from pyvirtualdisplay import Display
import os
display = Display(visible=0, size=(1280, 1024))
display.start()
from mayavi import mlab
def plot_to_buffer(fig):
""" Save figure to a virtual file. Used to append data to gif or mp4 in animate 2D functions.
"""
import io
import numpy as np
with io.Byt... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/raster/utils.py | 0.642993 | 0.487002 | utils.py | pypi |
from optparse import OptionParser
import rasterio
from affine import Affine # For easly manipulation of affine matrix
from rasterio.warp import reproject, Resampling, array_bounds, calculate_default_transform
import numpy as np
def summary(dataset):
print(" - no data value:", dataset.nodata )
print(" - transf... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/raster/rotate_and_rescale.py | 0.755276 | 0.550487 | rotate_and_rescale.py | pypi |
from optparse import OptionParser
def convert_size(size_bytes):
import math
if size_bytes == 0:
return "0B"
size_name = ("B", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB")
i = int(math.floor(math.log(size_bytes, 1024)))
p = math.pow(1024, i)
s = round(size_bytes / p, 2)
return "%... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/raster/interpolate.py | 0.672117 | 0.739681 | interpolate.py | pypi |
import requests
import os
import rasterio
from rasterio import mask as msk
import fiona
from shapely.geometry import shape, Point, Polygon
from optparse import OptionParser
from tqdm import tqdm
from osgeo import gdal
from pathlib import Path;
from os import walk
from os.path import exists
import re
import glob
def tr... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/chelsa/utils.py | 0.590425 | 0.444263 | utils.py | pypi |
def request(
inputFile=None,
variables=None,
timesID=None,
points=None,
margin=0.0,
world_dir='chelsa-world',
landscape_dir='chelsa-landscape',
geotiff='chelsa-stacked.tif',
cleanup=False
) -> None:
"""
Downloads bio and orog variables ... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/chelsa/request.py | 0.495117 | 0.339171 | request.py | pypi |
def get_variables_args(option, opt, value, parser):
"""
Helper function
"""
setattr(parser.values, option.dest, value.split(','))
def get_timesID_args(option, opt, value, parser, type='int'):
"""
Helper function
"""
setattr(parser.values, option.dest, [int(s) for s in value.split(',')]... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/chelsa/cli/options.py | 0.715225 | 0.315578 | options.py | pypi |
def paginated_search(max_limit, *args, **kwargs):
"""
In its current version, pygbif can not search more than 300 occurrences at
once: this solves a bit of the problem.
"""
MAX_LIMIT = max_limit
PER_PAGE = 100
results = []
from pygbif import occurrences
if(MAX_LIMIT <= PER_PAGE):
... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/gbif/utils.py | 0.641198 | 0.344278 | utils.py | pypi |
def request(scientific_name: str,
points: str,
buffer: float = 0.0,
all: bool = False,
limit: int = 30,
year = None,
csv_file: str = "occurrences.csv",
shapefile: str = "occurrences.shp",
) ->... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/gbif/request.py | 0.837919 | 0.558568 | request.py | pypi |
def get_parser():
"""
Returns parsed options from command line
"""
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-s", "--species",
type="str",
dest="scientific_name",
help="Species name for ... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/gbif/cli/options.py | 0.886101 | 0.289422 | options.py | pypi |
def get_variables_args(option, opt, value, parser):
"""
Helper function
"""
setattr(parser.values, option.dest, value.split(','))
def get_parser():
"""
Returns parsed options from command line
"""
from optparse import OptionParser
parser = OptionParser()
parser.add_option("-... | /quetzal_crumbs-1.0.17-py3-none-any.whl/crumbs/sdm/cli/fit_options.py | 0.855791 | 0.311924 | fit_options.py | pypi |
import re
import quetzpy.modules.data_cleaning as data_cleaning
def hashtag_extractor(tweet_text: str) -> list:
"""
Extract hashtags from a tweet\n
Returns a list of hashtags
"""
try:
# hashtags = pattern.match(tweet_text)
hashtags = re.findall(r"#(\w+)", tweet_text)
except:
... | /modules/data_mining.py | 0.716119 | 0.291529 | data_mining.py | pypi |
import re
def hashtags_remover(tweet_text: str) -> str:
"""
Remove the hashtags from a tweet\n
:tweet_text: text from a tweet\n
:return: tweet_text cleaned from hashtags
"""
pattern = re.compile(r"\s#+\S+")
tweet_text_cleaned = pattern.sub("", tweet_text)
return tweet_text_cleaned
de... | /modules/data_cleaning.py | 0.546254 | 0.213644 | data_cleaning.py | pypi |
import logging
from multiprocessing import JoinableQueue, Manager, Process, Queue
from multiprocessing.managers import SyncManager
from typing import Callable, Dict, Iterable, List, Sequence, Union
from .constants import QueueFlags, QueueNames
logger = logging.getLogger('QueueAutomator')
class QueueAutomator:
"... | /queue_automator-0.1.0-py3-none-any.whl/queue_automator/automator.py | 0.832169 | 0.152505 | automator.py | pypi |
from functools import lru_cache
from math import ceil
from os import cpu_count
from typing import Any, Callable, Iterable, List, Optional
from .automator import QueueAutomator
from .constants import QueueNames
class MaybeWrapper:
def __init__(self, func: Callable, nothing_check: Callable, default: Any = None) -... | /queue_automator-0.1.0-py3-none-any.whl/queue_automator/maybe.py | 0.87464 | 0.385433 | maybe.py | pypi |
from python_helper import Constant as c
from python_helper import ObjectHelper, StringHelper
from python_framework import Mapper, MapperMethod, StaticConverter
from constant import SubscriptionConstant
import SubscriptionDto
import SubscriptionModel
@Mapper()
class SubscriptionModelMapper:
@MapperMethod(reques... | /queue_manager_api-0.1.42.tar.gz/queue_manager_api-0.1.42/api/src/mapper/SubscriptionModelMapper.py | 0.443841 | 0.185615 | SubscriptionModelMapper.py | pypi |
import math
import os
import stat
import logging
import sys
import shutil
import time
def assign_tasks_to_chunks(num_tasks, num_chunks):
"""
When you have too many tasks for your parallel processing queue this
function chunks multiple tasks into fewer chunks.
Parameters
----------
num_tasks :... | /queue_map_reduce_sebastian_achim_mueller-1.1.1-py3-none-any.whl/queue_map_reduce/utils.py | 0.531209 | 0.342462 | utils.py | pypi |
def make_worker_node_script(func_module, func_name, environ):
"""
Returns a string that is a python-script.
This python-script will be executed on the worker-node.
In here, the environment variables are set explicitly.
It reads the chunk of tasks, runs result = func(task), and writes the
results... | /queue_map_reduce_sebastian_achim_mueller-1.1.1-py3-none-any.whl/queue_map_reduce/queue/worker_node_script.py | 0.552781 | 0.629148 | worker_node_script.py | pypi |
import pandas as pd
def fill_rate(df_enter: pd.DataFrame, df_leave: pd.DataFrame):
"""
Compute the cumulative fill rate over the entire dataset (this is not an average)
Assuming a wide formatted frame with index date as some resolution (e.g. week).
The columns of the frames will be used to calculate t... | /queue_metrics-0.0.1-py3-none-any.whl/queue_metrics/cumulative.py | 0.881468 | 0.818592 | cumulative.py | pypi |
# ==============================================================
# ssqs function
# ==============================================================
def ssqs(**parameters):
"""This function calculates the parameters of various single-server queueing systems.
A queueing system is considered stable when the arriv... | /queueing_systems-0.1.2.tar.gz/queueing_systems-0.1.2/queueing_systems/functions.py | 0.895406 | 0.651826 | functions.py | pypi |
Overview
========
Queueing-tool is a Python simulation package for analyzing networks of queues.
The simulations are event based, where events are comprised as arrivals and
departures of agents that move from queue to queue in the network. The network
is represented as a graph, which is handled by NetworkX.
There are... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/docs/overview.rst | 0.960759 | 0.868715 | overview.rst | pypi |
import functools
import numpy as np
import queueing_tool as qt
# Make an adjacency list
adja_list = {0: {1: {}}, 1: {k: {} for k in range(2, 22)}}
# Make an object that has the same dimensions as your adjacency list that
# specifies the type of queue along each edge.
edge_list = {0: {1: 1}, 1: {k: 2 for k in range(... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/examples/example_grocery_store.py | 0.761272 | 0.562357 | example_grocery_store.py | pypi |
class UnionFind(object):
"""The union-find data structure with union by rank and path compression.
The UnionFind data structure is a collection of objects that supports
the union and find operations (described below). Each object in the
collection belongs to a set, which is identified by its leader. Us... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/union_find.py | 0.90482 | 0.846197 | union_find.py | pypi |
import collections
import numbers
import copy
import array
import numpy as np
from numpy.random import uniform
try:
import matplotlib.pyplot as plt
from matplotlib.animation import FuncAnimation
from matplotlib.collections import LineCollection
plt.style.use('ggplot')
HAS_MATPLOTLIB = True
excep... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/network/queue_network.py | 0.815196 | 0.603348 | queue_network.py | pypi |
import networkx as nx
import numpy as np
try:
import matplotlib.pyplot as plt
from matplotlib.collections import LineCollection
plt.style.use('ggplot')
HAS_MATPLOTLIB = True
except ImportError:
HAS_MATPLOTLIB = False
def _matrix2dict(matrix, etype=False):
"""Takes an adjacency matrix and re... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/graph/graph_wrapper.py | 0.856872 | 0.59131 | graph_wrapper.py | pypi |
import networkx as nx
import numpy as np
from queueing_tool.graph.graph_functions import _test_graph, _calculate_distance
from queueing_tool.graph.graph_wrapper import (
adjacency2graph,
QueueNetworkDiGraph
)
def add_edge_lengths(g):
"""Add add the edge lengths as a :any:`DiGraph<networkx.DiGraph>`
f... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/graph/graph_preparation.py | 0.891014 | 0.72899 | graph_preparation.py | pypi |
import networkx as nx
import numpy as np
from queueing_tool.graph.graph_wrapper import QueueNetworkDiGraph
def _calculate_distance(latlon1, latlon2):
"""Calculates the distance between two points on earth.
"""
lat1, lon1 = latlon1
lat2, lon2 = latlon2
dlon = lon2 - lon1
dlat = lat2 - lat1
... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/graph/graph_functions.py | 0.887631 | 0.693668 | graph_functions.py | pypi |
from heapq import heappush, heappop
from numpy import logical_or, infty
import numpy as np
from queueing_tool.queues.agents import Agent
from queueing_tool.queues.queue_servers import LossQueue
class ResourceAgent(Agent):
"""An agent designed to interact with the :class:`.ResourceQueue`
class.
When an ... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/queues/queue_extentions.py | 0.847211 | 0.270227 | queue_extentions.py | pypi |
import collections
import copy
import numbers
from heapq import heappush, heappop
from numpy.random import uniform, exponential
from numpy import infty
import numpy as np
from queueing_tool.queues.agents import Agent, InftyAgent
def poisson_random_measure(t, rate, rate_max):
"""A function that returns the arriv... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/queues/queue_servers.py | 0.923108 | 0.776983 | queue_servers.py | pypi |
from numpy import infty
from numpy.random import uniform
from queueing_tool.queues.choice import _choice, _argmin
class Agent(object):
"""The base class for an agent.
``Agents`` are the objects that move throughout the network.
``Agents`` are instantiated by a queue, and once serviced the
``Agent`` ... | /queueing-tool-1.2.5.tar.gz/queueing-tool-1.2.5/queueing_tool/queues/agents.py | 0.951006 | 0.718168 | agents.py | pypi |
import csv
import dataclasses
import gzip
from io import StringIO
from typing import Any, Dict, Iterator, List, Literal, Optional, Union, overload
import requests
from queuery_client.cast import cast_row
try:
import pandas
except ModuleNotFoundError:
pandas = None
@dataclasses.dataclass
class ResponseBody:... | /queuery_client-1.0.2-py3-none-any.whl/queuery_client/response.py | 0.824603 | 0.206694 | response.py | pypi |
import math
import torch
import torch.nn as nn
from torch.nn.parameter import Parameter
from torch.autograd import Function
class BinarizeF(Function):
@staticmethod
def forward(cxt, input):
output = input.new(input.size())
output[input >= 0] = 1
output[input < 0] = -1
return ou... | /qsk_circ/QuantumFlow.py | 0.828211 | 0.531392 | QuantumFlow.py | pypi |
import numpy as np
import qiskit
import torch
from torch.autograd import Function
class QuantumCircuit:
"""
This class provides a simple interface for interaction
with the quantum circuit
"""
def __init__(self, n_qubits, backend, shots):
# --- Circuit definition ---
print('... | /qsk_circ/QuantumCircuit.py | 0.905348 | 0.656885 | QuantumCircuit.py | pypi |
from __future__ import annotations
# Core Imports
import time
from pathlib import Path
from string import Template
# Internal Imports
from quicfire_tools.parameters import SimulationParameters
class SimulationInputs:
"""
Input Module
"""
def __init__(self, directory: Path | str):
if isinsta... | /quicfire_tools-0.1.4-py3-none-any.whl/quicfire_tools/inputs.py | 0.853745 | 0.296031 | inputs.py | pypi |
import six
class OpenApiException(Exception):
"""The base exception class for all OpenAPIExceptions"""
class ApiTypeError(OpenApiException, TypeError):
def __init__(self, msg, path_to_item=None, valid_classes=None,
key_type=None):
""" Raises an exception for TypeErrors
Args... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick_client/exceptions.py | 0.696268 | 0.26309 | exceptions.py | pypi |
import argparse
import sys
from gettext import gettext
class QuickArgParser(argparse.ArgumentParser):
"""
Custom quick arg parser
The arg parser customizes the error message. It includes a pointer to the help argument.
"""
def __init__(self, *args, **kwargs):
super().__init__(*args, **k... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick/parser.py | 0.469034 | 0.173603 | parser.py | pypi |
import argparse
import dataclasses
from argparse import ArgumentParser
from argparse import _SubParsersAction
from typing import List
from typing import Optional
from quick.driver import create_quick_parser
@dataclasses.dataclass
class Argument:
name: str
help: Optional[str]
prog: Optional[str]
class ... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick/generate_docs.py | 0.763087 | 0.206674 | generate_docs.py | pypi |
from argparse import ArgumentParser
import isodate
from quick_client import ApiException
from quick_client import GatewaySchema
from quick_client.models import QuickTopicType
from quick_client.models import TopicCreationData
from quick_client.models import TopicWriteType
from quick.commands.base import ArgumentGroup... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick/commands/topic.py | 0.476823 | 0.186132 | topic.py | pypi |
from argparse import ArgumentParser
from quick_client import ApiException
from quick_client import MirrorCreationData
from quick.commands.base import ArgumentGroup
from quick.commands.base import Group
from quick.commands.base import ManagerCommand
class CreateMirror(ManagerCommand):
name = "create"
help = ... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick/commands/mirror.py | 0.758242 | 0.156878 | mirror.py | pypi |
from argparse import ArgumentParser
from argparse import FileType
from quick_client import ApiException
from quick_client import GatewayCreationData
from quick_client import SchemaData
from quick.commands.base import ArgumentGroup
from quick.commands.base import Group
from quick.commands.base import ManagerCommand
... | /quick-cli-0.8.0.tar.gz/quick-cli-0.8.0/quick/commands/gateway.py | 0.723602 | 0.158142 | gateway.py | pypi |
import os
from typing import Dict, AnyStr, Union, Any, List
import logging
from quick_config.utils.module_loader import load_env_var_by_file
from quick_config.constants import ENV_KEY, TEST_ALIASES, STG_ALIASES, PROD_ALIASES, DEV_ALIASES, BASE_LOG_FMT
class _Config:
def __init__(self, env_path, environment_key=EN... | /quick_config-0.2.4-py3-none-any.whl/quick_config/provider.py | 0.665628 | 0.186336 | provider.py | pypi |
from scrapy import signals
# useful for handling different item types with a single interface
from itemadapter import is_item, ItemAdapter
class NewsSiteSpiderMiddleware:
# Not all methods need to be defined. If a method is not defined,
# scrapy acts as if the spider middleware does not modify the
# pas... | /quick_crawler-0.0.8-py3-none-any.whl/quick_crawler/scrapy_projects/news_site/news_site/middlewares.py | 0.68637 | 0.270967 | middlewares.py | pypi |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.