inputs
stringlengths
312
52k
targets
stringlengths
1
3.1k
block_type
stringclasses
11 values
scenario
stringclasses
7 values
<filename>microagents/agents/agent_similarity.py<fim_prefix>import logging import numpy as np from typing import List, Tuple, Optional from sklearn.metrics.pairwise import cosine_similarity from integrations.openaiwrapper import OpenAIAPIWrapper logger = logging.getLogger() class Agent: def __init__(self, purpos...
if similarity > highest_similarity: highest_similarity = similarity closest_agent = agent
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/agents/agent_similarity.py<fim_prefix>import logging import numpy as np from typing import List, Tuple, Optional from sklearn.metrics.pairwise import cosine_similarity from integrations.openaiwrapper import OpenAIAPIWrapper logger = logging.getLogger() class Agent: def __init__(self, purpos...
if agent.purpose_embedding is None: agent.purpose_embedding = self.get_embedding(agent.purpose)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/agents/agent_similarity.py<fim_prefix>import logging import numpy as np from typing import List, Tuple, Optional from sklearn.metrics.pairwise import cosine_similarity from integrations.openaiwrapper import OpenAIAPIWrapper logger = logging.getLogger() class Agent: def __init__(self, purpos...
if 'data' in response and len(response['data']) > 0 and 'embedding' in response['data'][0]: return np.array(response['data'][0]['embedding']) else: logger.exception("Invalid response format") raise ValueError("Invalid response format")
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/integrations/memoize.py<fim_prefix>import sqlite3 import hashlib import json import functools ## Originally from https://www.kevinkatz.io/posts/memoize-to-sqlite def memoize_to_sqlite(func_name: str, filename: str = "cache.db"): """ Memoization decorator that caches the output of a metho...
if result is not None: return result
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/agents/agent_persistence_manager.py<fim_prefix>from agents.agent_serializer import AgentSerializer from integrations.memoize import memoize_to_sqlite from integrations.sqlite_agent_persistence import SQLiteAgentPersistence class AgentPersistenceManager: def __init__(self, db_filename="agents...
if agent: agents.append(agent)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/agents/microagent.py<fim_prefix>import logging import uuid from integrations.openaiwrapper import OpenAIAPIWrapper from agents.agent_evaluation import AgentEvaluator from agents.agent_response import AgentResponse from agents.agent_similarity import AgentSimilarity from agents.response_extraction ...
if parent: self.parent_id = parent.id if parent else None else: self.parent_id = None
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/integrations/memoize.py<fim_prefix>import sqlite3 import hashlib import json import functools ## Originally from https://www.kevinkatz.io/posts/memoize-to-sqlite def memoize_to_sqlite(func_name: str, filename: str = "cache.db"): """ Memoization decorator that caches the output of a metho...
def decorator(func): @functools.wraps(func) def wrapped(*args, **kwargs): with SQLiteMemoization(filename) as memoizer: return memoizer.fetch_or_compute(func, func_name, *args, **kwargs) return wrapped
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/integrations/memoize.py<fim_prefix>import sqlite3 import hashlib import json import functools ## Originally from https://www.kevinkatz.io/posts/memoize-to-sqlite def memoize_to_sqlite(func_name: str, filename: str = "cache.db"): """ Memoization decorator that caches the output of a metho...
def wrapped(*args, **kwargs): with SQLiteMemoization(filename) as memoizer: return memoizer.fetch_or_compute(func, func_name, *args, **kwargs)
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/integrations/memoize.py<fim_prefix>import sqlite3 import hashlib import json import functools ## Originally from https://www.kevinkatz.io/posts/memoize-to-sqlite def memoize_to_sqlite(func_name: str, filename: str = "cache.db"): """ Memoization decorator that caches the output of a metho...
def wrapped(*args, **kwargs): with SQLiteMemoization(filename) as memoizer: return memoizer.fetch_or_compute(func, func_name, *args, **kwargs)
METHOD
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/integrations/sqlite_agent_persistence.py<fim_prefix>import sqlite3 import json from integrations.agent_persistence import AbstractAgentPersistence class SQLiteAgentPersistence(AbstractAgentPersistence): def __init__(self, filename="agents.db"): self.filename = filename self._i...
# add id field
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microagents/agents/microagent.py<fim_prefix>import logging import uuid from integrations.openaiwrapper import OpenAIAPIWrapper from agents.agent_evaluation import AgentEvaluator from agents.agent_response import AgentResponse from agents.agent_similarity import AgentSimilarity from agents.response_extraction ...
# Initialize components used by the agent
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
result[url] = idf_score * numerator / denominator
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: <fim_suffix> else: old[url] =...
old[url] += score
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
idf_score = self.idf(kw)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
self._index: dict[str, dict[str, int]] = defaultdict(lambda: defaultdict(int))
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
self._documents[url] = content
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
denominator = freq + self.k1 * ( 1 - self.b + self.b * len(self._documents[url]) / avdl )
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
result = {}
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
string_without_double_spaces = " ".join(string_without_punc.split())
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
return url_scores
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
n_kw = len(self.get_urls(kw))
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
for word in words: self._index[word][url] += 1
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): <fim_suffix> return old def normalize_string(input_string: str) -> str: translation_table = str.maketrans...
for url, score in new.items(): if url in old: old[url] += score else: old[url] = score
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
for kw in keywords: kw_urls_score = self.bm25(kw) url_scores = update_url_scores(url_scores, kw_urls_score)
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
for url, freq in self.get_urls(kw).items(): numerator = freq * (self.k1 + 1) denominator = freq + self.k1 * ( 1 - self.b + self.b * len(self._documents[url]) / avdl ) result[url] = idf_score * numerator / denominator
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): if url in old: old[url] += score else: old[u...
for url, content in documents: self.index(url, content)
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>microsearch/src/microsearch/engine.py<fim_prefix>from collections import defaultdict from math import log import string def update_url_scores(old: dict[str, float], new: dict[str, float]): for url, score in new.items(): <fim_suffix> return old def normalize_string(input_string: str) -> str...
if url in old: old[url] += score else: old[url] = score
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
self.list_type = list_types[text[0]]
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
line_set.add(line_without_numbers)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
block = { "block_idx": block_idx, "block_text": running_line, "block_type": line_type, "text_group_start_idx": -1, "block_list": [], "header_block_idx": header_block_idx, "level": ...
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
self.length = len(self.text)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
n = self.check_numeric()
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
self.text = token
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
modified_sents = nltk_tokenzier.tokenize(modified_text)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
self.is_table_row = False
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/spell_utils.py<fim_prefix>import logging import os import string from symspellpy.symspellpy import SymSpell from symspellpy.symspellpy import Verbosity import nlm_ingestor.ingestor as ingestor from nlm_ingestor.ingestor import patterns logger = logging.getLogger(__n...
suggestion = self.sym_spell.word_segmentation(input_term)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
return sents
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
if len(line_buffer) > 0: # find out if previous line was a discontinous line prev_line = line_buffer[-1] logger.debug("========") logger.debug(f"{prev_line.incomplete_line} >> {prev_line.text} \n") logger.debug(f"{curr_line.continuing_line} >> {curr_line.tex...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
if not line_type == "list_item": line_type = "para"
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/spell_utils.py<fim_prefix>import logging import os import string from symspellpy.symspellpy import SymSpell from symspellpy.symspellpy import Verbosity import nlm_ingestor.ingestor as ingestor from nlm_ingestor.ingestor import patterns logger = logging.getLogger(__n...
if is_mixed_case_term: corrected_string = string.capwords(corrected_string)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
if n: n = round(float(n)) if n > 0: digits = int(math.log10(n)) + 1 elif n == 0: digits = 1 else: digits = int(math.log10(-n)) + 2 self.num_digits = digits if digit...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
if ( is_incomplete and not (curr_line.is_list_or_row or curr_line.line_type == "list_item") ) or curr_line.continuing_line: logger.debug("connecting..") running_line = formatter.connect(running_line, curr_line.text) line_buffer....
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
if n > 0: digits = int(math.log10(n)) + 1 elif n == 0: digits = 1 else: digits = int(math.log10(-n)) + 2
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
if not self.is_header: value_count = ( self.number_count + self.dollar_count + self.pct_count + self.text.count(" - ") ) word_symbols = self.word_count - self.dollar_sign_count if word_symbols == 0: ...
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
for org_text in org_texts.split("\n"): org_text = space_rule.sub(r'\1', org_text) modified_text = re.sub(r'^([.,?!]\s+)+', "", org_text) # To handle bug https://github.com/nltk/nltk/issues/2925 orig_offset = abs(len(org_text) - len(modified_text)) # do not break bracket for spa...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
for rule, replaced in rules: modified_text = rule.sub(replaced, modified_text)
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
for line_str in lines: # print(line_str) line_str = clean_line(line_str) if should_skip(line_str, xml=xml): continue line_without_numbers = re.sub(r"\d+", "", line_str) if line_without_numbers in line_set: continue else: line_set.add(l...
FOR
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
# if line_type == "list_item":
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
# if len(span.split()) >= 10:
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
# year does not count as a number
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
# edge case for html and markdown
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
# cut org_text based on lengths of modified_sent
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
# commit the line and start a new line
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
# skip this logic when span is too big? disabled for now
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
# if we are connecting lines, then this has to be a para unless it is a list_item, basically no headers
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/processors.py<fim_prefix>import logging import re from collections import Counter from collections import defaultdict from . import formatter from . import line_parser from . import patterns from nlm_ingestor.ingestor_utils import spell_utils from nlm_ingestor.ingestor_util...
# running_line = running_line[1:].lstrip()
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
# do not break bracket
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor_utils/utils.py<fim_prefix>import json import re import numpy as np from nltk import load from nltk import PunktSentenceTokenizer nltk_abbs = load("tokenizers/punkt/{}.pickle".format("english"))._params.abbrev_types class NpEncoder(json.JSONEncoder): def default(self...
while offset < len(modified_text) and sent_idx < len(modified_sents): if modified_text[offset] == " ": offset += 1 continue # cut org_text based on lengths of modified_sent modified_sent = modified_sents[sent_idx] sents.append(org_text[off...
WHILE
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
try: if n: n = round(float(n)) if n > 0: digits = int(math.log10(n)) + 1 elif n == 0: digits = 1 else: digits = int(math.log10(-n)) + 2 self.num_digits = digits ...
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>nlm-ingestor/nlm_ingestor/ingestor/line_parser.py<fim_prefix>import datetime import logging import math import re import string from nltk.corpus import stopwords from .patterns import abbreviations from .patterns import states from .patterns import states_abbreviations from .styling_utils import mode_of_lis...
except Exception as e: logging.error(e) self.num_digits = 0
CATCH
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if (realpath := _as_path(path)) and realpath.exists(): modelfile = self._parse_modelfile(realpath.read_text(), base=realpath.parent) elif modelfile: modelfile = self._parse_modelfile(modelfile) else: raise RequestError('must provide either path or modelfile')
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if stream: return await self._stream(*args, **kwargs)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if path.exists(): args = f'@{await self._create_blob(path)}\n'
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if command.upper() not in ['FROM', 'ADAPTER']: print(line, end='', file=out) continue
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if b := _as_bytesio(image): return b64encode(b.read()).decode('utf-8')
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if command.upper() not in ['FROM', 'ADAPTER']: print(line, end='', file=out) continue
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if (realpath := _as_path(path)) and realpath.exists(): modelfile = await self._parse_modelfile(realpath.read_text(), base=realpath.parent) elif modelfile: modelfile = await self._parse_modelfile(modelfile) else: raise RequestError('must provide either path or modelfile')
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if not chunk: break
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if not chunk: break
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
if isinstance(s, io.BytesIO): return s elif isinstance(s, bytes): return io.BytesIO(s)
IF
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
print(command, args, end='', file=out)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
path = path if path.is_absolute() else base / path
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
return b64encode(b.read()).decode('utf-8')
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
b64decode(image, validate=True)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
command, _, args = line.partition(' ')
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
return self._stream(*args, **kwargs) if stream else self._request(*args, **kwargs).json()
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
modelfile = self._parse_modelfile(modelfile)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_types.py<fim_prefix>import json from typing import Any, TypedDict, Sequence, Literal import sys if sys.version_info < (3, 11): from typing_extensions import NotRequired else: from typing import NotRequired class BaseGenerateResponse(TypedDict): model: str 'Model used to gener...
error = json.loads(error).get('error', error)
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
split = urllib.parse.urlsplit('://'.join([scheme, hostport]))
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
yield partial
STATEMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" >>> _parse_host(None) 'http://127.0.0.1:11434' >>> _parse_host('') 'http://127.0.0.1:11434' >>> _parse_host('1.2.3.4') 'http://1.2.3.4:11434' >>> _parse_host(':56789') 'http://127.0.0.1:56789' >>> _parse_host('1.2.3.4:56789') 'http://1.2.3.4:56789' >>> _parse_host('http://1.2.3.4') 'http://1...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Raises `ResponseError` if the request could not be fulfilled. Returns `ProgressResponse` if `stream` is `False`, otherwise returns a `ProgressResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Create a chat response using the requested model. Raises `RequestError` if a model is not provided. Raises `ResponseError` if the request could not be fulfilled. Returns `ChatResponse` if `stream` is `False`, otherwise returns an asynchronous `ChatResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Raises `ResponseError` if the request could not be fulfilled. Returns `ProgressResponse` if `stream` is `False`, otherwise returns a `ProgressResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Create a response using the requested model. Raises `RequestError` if a model is not provided. Raises `ResponseError` if the request could not be fulfilled. Returns `GenerateResponse` if `stream` is `False`, otherwise returns a `GenerateResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Raises `ResponseError` if the request could not be fulfilled. Returns `ProgressResponse` if `stream` is `False`, otherwise returns a `ProgressResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Raises `ResponseError` if the request could not be fulfilled. Returns `ProgressResponse` if `stream` is `False`, otherwise returns a `ProgressResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Creates a httpx client. Default parameters are the same as those defined in httpx except for the following: - `follow_redirects`: True - `timeout`: None `kwargs` are passed to the httpx client. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" >>> _encode_image(b'ollama') 'b2xsYW1h' >>> _encode_image(io.BytesIO(b'ollama')) 'b2xsYW1h' >>> _encode_image('LICENSE') 'TUlUIExpY2Vuc2UKCkNvcHlyaWdodCAoYykgT2xsYW1hCgpQZXJtaXNzaW9uIGlzIGhlcmVieSBncmFudGVkLCBmcmVlIG9mIGNoYXJnZSwgdG8gYW55IHBlcnNvbiBvYnRhaW5pbmcgYSBjb3B5Cm9mIHRoaXMgc29mdHdhcmUgYW5kIGFzc2...
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
""" Raises `ResponseError` if the request could not be fulfilled. Returns `ProgressResponse` if `stream` is `False`, otherwise returns a `ProgressResponse` generator. """
BLOCK_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: r.raise_for_status()
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: await self._request('HEAD', f'/api/blobs/{digest}')
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: self._request('HEAD', f'/api/blobs/{digest}')
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: r.raise_for_status()
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: r.raise_for_status()
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_types.py<fim_prefix>import json from typing import Any, TypedDict, Sequence, Literal import sys if sys.version_info < (3, 11): from typing_extensions import NotRequired else: from typing import NotRequired class BaseGenerateResponse(TypedDict): model: str 'Model used to gener...
try: # try to parse content as JSON and extract 'error' # fallback to raw content if JSON parsing fails error = json.loads(error).get('error', error)
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
try: b64decode(image, validate=True) return image if isinstance(image, str) else image.decode('utf-8')
TRY
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_types.py<fim_prefix>import json from typing import Any, TypedDict, Sequence, Literal import sys if sys.version_info < (3, 11): from typing_extensions import NotRequired else: from typing import NotRequired class BaseGenerateResponse(TypedDict): model: str 'Model used to gener...
# try to parse content as JSON and extract 'error'
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_types.py<fim_prefix>import json from typing import Any, TypedDict, Sequence, Literal import sys if sys.version_info < (3, 11): from typing_extensions import NotRequired else: from typing import NotRequired class BaseGenerateResponse(TypedDict): model: str 'Model used to gener...
# fallback to raw content if JSON parsing fails
LINE_COMMENT
prefix_suffix_full_complete_current_block_no_evidence
<filename>ollama-python/ollama/_client.py<fim_prefix>import os import io import json import httpx import binascii import platform import urllib.parse from os import PathLike from pathlib import Path from hashlib import sha256 from base64 import b64encode, b64decode from typing import Any, AnyStr, Union, Optional, Sequ...
except (binascii.Error, TypeError): ...
CATCH
prefix_suffix_full_complete_current_block_no_evidence