| """Contains useful utility functions.""" |
|
|
| import json |
| import urllib.parse |
| from pathlib import Path |
|
|
| import requests |
| import tiktoken |
| from openai import AsyncOpenAI, OpenAI |
| from openai.types.beta.threads.run import Run |
|
|
| from virtual_lab.constants import ( |
| DEFAULT_FINETUNING_EPOCHS, |
| MODEL_TO_INPUT_PRICE_PER_TOKEN, |
| MODEL_TO_OUTPUT_PRICE_PER_TOKEN, |
| FINETUNING_MODEL_TO_TRAINING_PRICE_PER_TOKEN, |
| PUBMED_TOOL_NAME, |
| ) |
| from virtual_lab.prompts import format_references |
|
|
|
|
| def get_pubmed_central_article( |
| pmcid: str, abstract_only: bool = False |
| ) -> tuple[str | None, list[str] | None]: |
| """Gets the title and content (abstract or full text) of a PubMed Central article given a PMC ID. |
| |
| Note: This only returns main text, ignoring tables, figures, and references. |
| |
| :param pmcid: The PMC ID of the article. |
| :param abstract_only: Whether to return only the abstract instead of the full text. |
| :return: The title and content (abstract or full text of the article as a list of paragraphs) |
| or None if the article is not found. |
| """ |
| |
| text_url = f"https://www.ncbi.nlm.nih.gov/research/bionlp/RESTful/pmcoa.cgi/BioC_JSON/PMC{pmcid}/unicode" |
| response = requests.get(text_url) |
| response.raise_for_status() |
|
|
| |
| try: |
| article = response.json() |
| except json.JSONDecodeError: |
| return None, None |
|
|
| |
| document = article[0]["documents"][0] |
|
|
| |
| title = next( |
| passage["text"] |
| for passage in document["passages"] |
| if passage["infons"]["section_type"] == "TITLE" |
| ) |
|
|
| |
| passages = [ |
| passage |
| for passage in document["passages"] |
| if passage["infons"]["type"] in {"abstract", "paragraph"} |
| ] |
|
|
| |
| if abstract_only: |
| passages = [ |
| passage |
| for passage in passages |
| if passage["infons"]["section_type"] in ["ABSTRACT"] |
| ] |
| else: |
| passages = [ |
| passage |
| for passage in passages |
| if passage["infons"]["section_type"] |
| in ["ABSTRACT", "INTRO", "RESULTS", "DISCUSS", "CONCL", "METHODS"] |
| ] |
|
|
| |
| content = [passage["text"] for passage in passages] |
|
|
| return title, content |
|
|
|
|
| def run_pubmed_search( |
| query: str, num_articles: int = 3, abstract_only: bool = False |
| ) -> str: |
| """Runs a PubMed search, returning the full text of the top matching article. |
| |
| :param query: The query to search PubMed with. |
| :param num_articles: The number of articles to search for. |
| :param abstract_only: Whether to return only the abstract instead of the full text. |
| :return: The full text of the top matching article. |
| """ |
| |
| print( |
| f'Searching PubMed Central for {num_articles} articles ({'abstracts' if abstract_only else 'full text'}) with query: "{query}"' |
| ) |
|
|
| |
| search_url = f"https://eutils.ncbi.nlm.nih.gov/entrez/eutils/esearch.fcgi?db=pmc&term={urllib.parse.quote_plus(query)}&retmax={2 * num_articles}&retmode=json&sort=relevance" |
| response = requests.get(search_url) |
| response.raise_for_status() |
| pmcids_found = response.json()["esearchresult"]["idlist"] |
|
|
| |
| texts = [] |
| titles = [] |
| pmcids = [] |
|
|
| for pmcid in pmcids_found: |
| |
| if len(pmcids) >= num_articles: |
| break |
|
|
| title, content = get_pubmed_central_article( |
| pmcid=pmcid, |
| abstract_only=abstract_only, |
| ) |
|
|
| if title is None: |
| continue |
|
|
| texts.append(f"PMCID = {pmcid}\n\nTitle = {title}\n\n{'\n\n'.join(content)}") |
| titles.append(title) |
| pmcids.append(pmcid) |
|
|
| |
| article_count = len(texts) |
|
|
| print(f"Found {article_count:,} articles on PubMed Central") |
|
|
| |
| if article_count == 0: |
| combined_text = f'No articles found on PubMed Central for the query "{query}".' |
| else: |
| combined_text = format_references( |
| references=tuple(texts), |
| reference_type="paper", |
| intro=f'Here are the top {article_count} articles on PubMed Central for the query "{query}":', |
| ) |
|
|
| return combined_text |
|
|
|
|
| def run_tools(run: Run) -> list[dict[str, str]]: |
| """Runs the tools in a required action. |
| |
| :param run: The run to run tools for. |
| :return: A list of tool outputs. |
| """ |
| |
| tool_outputs = [] |
|
|
| |
| for tool in run.required_action.submit_tool_outputs.tool_calls: |
| if tool.function.name == PUBMED_TOOL_NAME: |
| |
| args_dict = json.loads(tool.function.arguments) |
|
|
| |
| tool_outputs.append( |
| { |
| "tool_call_id": tool.id, |
| "output": run_pubmed_search(**args_dict), |
| } |
| ) |
| else: |
| raise ValueError(f"Unknown tool: {tool.function.name}") |
|
|
| return tool_outputs |
|
|
|
|
| def get_messages(client: OpenAI, thread_id: str) -> list[dict]: |
| """Gets messages from a thread. |
| |
| :param client: The OpenAI client. |
| :param thread_id: The ID of the thread to get messages from. |
| :return: A list of messages. |
| """ |
| |
| messages = [] |
| last_message = None |
| params = { |
| "thread_id": thread_id, |
| "limit": 100, |
| "order": "asc", |
| } |
|
|
| |
| while True: |
| |
| if last_message is not None: |
| params["after"] = last_message["id"] |
| elif "after" in params: |
| del params["after"] |
|
|
| |
| new_messages = [ |
| message.to_dict() for message in client.beta.threads.messages.list(**params) |
| ] |
|
|
| |
| messages += new_messages |
|
|
| |
| if len(new_messages) < params["limit"]: |
| break |
|
|
| |
| last_message = messages[-1] |
|
|
| |
| assert all(len(message["content"]) == 1 for message in messages) |
|
|
| return messages |
|
|
|
|
| async def async_get_messages(client: AsyncOpenAI, thread_id: str) -> list[dict]: |
| """Gets messages from a thread. |
| |
| :param client: The async OpenAI client. |
| :param thread_id: The ID of the thread to get messages from. |
| :return: A list of messages. |
| """ |
| |
| messages = [] |
| last_message = None |
| params = { |
| "thread_id": thread_id, |
| "limit": 100, |
| "order": "asc", |
| } |
|
|
| |
| while True: |
| |
| if last_message is not None: |
| params["after"] = last_message["id"] |
| elif "after" in params: |
| del params["after"] |
|
|
| |
| new_messages = [ |
| message.to_dict() |
| async for message in client.beta.threads.messages.list(**params) |
| ] |
|
|
| |
| messages += new_messages |
|
|
| |
| if len(new_messages) < params["limit"]: |
| break |
|
|
| |
| last_message = messages[-1] |
|
|
| |
| assert all(len(message["content"]) == 1 for message in messages) |
|
|
| return messages |
|
|
|
|
| def count_tokens(string: str, encoding_name: str = "cl100k_base") -> int: |
| """Returns the number of tokens in a text string. |
| |
| :param string: The text string to count tokens in. |
| :param encoding_name: The name of the encoding to use. |
| :return: The number of tokens in the text string. |
| """ |
| encoding = tiktoken.get_encoding(encoding_name) |
| num_tokens = len(encoding.encode(string)) |
|
|
| return num_tokens |
|
|
|
|
| def update_token_counts( |
| token_counts: dict[str, int], |
| discussion: list[dict[str, str]], |
| response: str, |
| ) -> None: |
| """Updates the token counts (in place) with a discussion and response. |
| |
| :param token_counts: The token counts to update. |
| :param discussion: The discussion to update the token counts with. |
| :param response: The response to update the token counts with. |
| """ |
| new_input_token_count = sum(count_tokens(turn["message"]) for turn in discussion) |
| new_output_token_count = count_tokens(response) |
|
|
| token_counts["input"] += new_input_token_count |
| token_counts["output"] += new_output_token_count |
|
|
| token_counts["max"] = max( |
| token_counts["max"], new_input_token_count + new_output_token_count |
| ) |
|
|
|
|
| def count_discussion_tokens( |
| discussion: list[dict[str, str]], |
| ) -> dict[str, int]: |
| """Counts the number of tokens in a discussion. |
| |
| :param discussion: The discussion to count tokens in. |
| :return: A dictionary of token counts. |
| """ |
| token_counts = { |
| "input": 0, |
| "output": 0, |
| "max": 0, |
| } |
|
|
| for index, turn in enumerate(discussion): |
| if turn["agent"] != "User": |
| update_token_counts( |
| token_counts=token_counts, |
| discussion=discussion[:index], |
| response=turn["message"], |
| ) |
|
|
| return token_counts |
|
|
|
|
| def compute_token_cost( |
| model: str, input_token_count: int, output_token_count: int |
| ) -> float: |
| """Computes the token cost of a model given input and output token counts. |
| |
| :param model: The name of the model. |
| :param input_token_count: The number of tokens in the input. |
| :param output_token_count: The number of tokens in the output. |
| :return: The token cost of the model. |
| """ |
| if ( |
| model not in MODEL_TO_INPUT_PRICE_PER_TOKEN |
| or model not in MODEL_TO_OUTPUT_PRICE_PER_TOKEN |
| ): |
| raise ValueError(f'Cost of model "{model}" not known') |
|
|
| return ( |
| input_token_count * MODEL_TO_INPUT_PRICE_PER_TOKEN[model] |
| + output_token_count * MODEL_TO_OUTPUT_PRICE_PER_TOKEN[model] |
| ) |
|
|
|
|
| def print_cost_and_time( |
| token_counts: dict[str, int], |
| model: str, |
| elapsed_time: float, |
| ) -> None: |
| |
| print(f"Input token count: {token_counts['input']:,}") |
| print(f"Output token count: {token_counts['output']:,}") |
| print(f"Tool token count: {token_counts['tool']:,}") |
| print(f"Max token length: {token_counts['max']:,}") |
|
|
| |
| try: |
| cost = compute_token_cost( |
| model=model, |
| input_token_count=token_counts["input"] + token_counts["tool"], |
| output_token_count=token_counts["output"], |
| ) |
| print(f"Cost: ${cost:.2f}") |
| except ValueError as e: |
| print(f"Warning: {e}") |
|
|
| |
| print(f"Time: {int(elapsed_time // 60)}:{int(elapsed_time % 60):02d}") |
|
|
|
|
| def compute_finetuning_cost( |
| model: str, token_count: int, num_epochs: int = DEFAULT_FINETUNING_EPOCHS |
| ) -> float: |
| """Computes the cost of fine-tuning a model. |
| |
| :param model: The model that will be finetuned. |
| :param token_count: The number of training tokens for finetuning. |
| :param num_epochs: Number of finetuning epochs. |
| :return: The cost of finetuning. |
| """ |
| if model not in FINETUNING_MODEL_TO_TRAINING_PRICE_PER_TOKEN: |
| raise ValueError(f'Cost of model "{model}" not known') |
|
|
| return ( |
| token_count * FINETUNING_MODEL_TO_TRAINING_PRICE_PER_TOKEN[model] * num_epochs |
| ) |
|
|
|
|
| def convert_messages_to_discussion( |
| messages: list[dict], assistant_id_to_title: dict[str, str] |
| ) -> list[dict[str, str]]: |
| """Converts OpenAI messages into discussion format (list of message dictionaries). |
| |
| :param messages: The messages to convert. |
| :param assistant_id_to_title: A dictionary mapping assistant IDs to titles. |
| :return: The discussion format (list of message dictionaries). |
| """ |
| return [ |
| { |
| "agent": ( |
| assistant_id_to_title[message["assistant_id"]] |
| if message["assistant_id"] is not None |
| else "User" |
| ), |
| "message": message["content"][0]["text"]["value"], |
| } |
| for message in messages |
| ] |
|
|
|
|
| def get_summary(discussion: list[dict[str, str]]) -> str: |
| """Get the summary from a discussion. |
| |
| :param discussion: The discussion to extract the summary from. |
| :return: The summary. |
| """ |
| return discussion[-1]["message"] |
|
|
|
|
| def load_summaries(discussion_paths: list[Path]) -> tuple[str, ...]: |
| """Load summaries from a list of discussion paths. |
| |
| :param discussion_paths: The paths to the discussion JSON files. The summary is the last entry in the discussion. |
| :return: A tuple of summaries. |
| """ |
| summaries = [] |
| for discussion_path in discussion_paths: |
| with open(discussion_path, "r") as file: |
| discussion = json.load(file) |
| summaries.append(get_summary(discussion)) |
|
|
| return tuple(summaries) |
|
|
|
|
| def save_meeting( |
| save_dir: Path, save_name: str, discussion: list[dict[str, str]] |
| ) -> None: |
| """Save a meeting discussion to JSON and Markdown files. |
| |
| :param save_dir: The directory to save the discussion. |
| :param save_name: The name of the discussion file that will be saved. |
| :param discussion: The discussion to save. |
| """ |
| |
| save_dir.mkdir(parents=True, exist_ok=True) |
|
|
| |
| with open(save_dir / f"{save_name}.json", "w") as f: |
| json.dump(discussion, f, indent=4) |
|
|
| |
| with open(save_dir / f"{save_name}.md", "w") as file: |
| for turn in discussion: |
| file.write(f"## {turn['agent']}\n\n{turn['message']}\n\n") |
|
|