code
stringlengths
161
233k
apis
listlengths
1
24
extract_api
stringlengths
162
68.5k
import time, ast, requests, warnings import numpy as np from llama_index import Document, ServiceContext, VectorStoreIndex from llama_index.storage.storage_context import StorageContext from llama_index.vector_stores import MilvusVectorStore from llama_index.node_parser import SentenceWindowNodeParser, HierarchicalNo...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.storage.storage_context.StorageContext.from_defaults", "llama_index.node_parser.HierarchicalNodeParser.from_defaults", "llama_index.node_parser.SentenceWindowNodeParser.from_defaults", "llama_index.ServiceContext.from_defaults", "llama_index.vect...
[((484, 517), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (507, 517), False, 'import time, ast, requests, warnings\n'), ((611, 743), 'llama_index.node_parser.SentenceWindowNodeParser.from_defaults', 'SentenceWindowNodeParser.from_defaults', ([], {'window_size': '(5)', '...
"""Llama Dataset Class.""" import asyncio import time from typing import List, Optional from llama_index.core.base.base_query_engine import BaseQueryEngine from llama_index.core.bridge.pydantic import Field from llama_index.core.llama_dataset.base import ( BaseLlamaDataExample, BaseLlamaDataset, BaseLlama...
[ "llama_index.core.bridge.pydantic.Field" ]
[((764, 909), 'llama_index.core.bridge.pydantic.Field', 'Field', ([], {'default_factory': 'str', 'description': '"""The generated (predicted) response that can be compared to a reference (ground-truth) answer."""'}), "(default_factory=str, description=\n 'The generated (predicted) response that can be compared to a ...
from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseGen, MessageRole, ) from llama_index.core.types import TokenGen def response_gen_from_query_engine(response_gen: TokenGen) -> ChatResponseGen: response_str = "" for token in response_gen: response_str...
[ "llama_index.core.base.llms.types.ChatMessage" ]
[((378, 439), 'llama_index.core.base.llms.types.ChatMessage', 'ChatMessage', ([], {'role': 'MessageRole.ASSISTANT', 'content': 'response_str'}), '(role=MessageRole.ASSISTANT, content=response_str)\n', (389, 439), False, 'from llama_index.core.base.llms.types import ChatMessage, ChatResponse, ChatResponseGen, MessageRol...
from typing import Dict, Any import asyncio # Create a new event loop loop = asyncio.new_event_loop() # Set the event loop as the current event loop asyncio.set_event_loop(loop) from llama_index import ( VectorStoreIndex, ServiceContext, download_loader, ) from llama_index.llama_pack.base import BaseLlam...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.llms.OpenAI", "llama_index.download_loader" ]
[((78, 102), 'asyncio.new_event_loop', 'asyncio.new_event_loop', ([], {}), '()\n', (100, 102), False, 'import asyncio\n'), ((151, 179), 'asyncio.set_event_loop', 'asyncio.set_event_loop', (['loop'], {}), '(loop)\n', (173, 179), False, 'import asyncio\n'), ((420, 607), 'streamlit.set_page_config', 'st.set_page_config', ...
"""DashScope llm api.""" from http import HTTPStatus from typing import Any, Dict, List, Optional, Sequence, Tuple from llama_index.legacy.bridge.pydantic import Field from llama_index.legacy.callbacks import CallbackManager from llama_index.legacy.constants import DEFAULT_NUM_OUTPUTS, DEFAULT_TEMPERATURE from llama_...
[ "llama_index.legacy.llms.base.llm_chat_callback", "llama_index.legacy.core.llms.types.ChatMessage", "llama_index.legacy.llms.base.llm_completion_callback", "llama_index.legacy.core.llms.types.LLMMetadata", "llama_index.legacy.llms.dashscope_utils.dashscope_response_to_chat_response", "llama_index.legacy.b...
[((2272, 2350), 'dashscope.Generation.call', 'Generation.call', ([], {'model': 'model', 'messages': 'messages', 'api_key': 'api_key'}), '(model=model, messages=messages, api_key=api_key, **parameters)\n', (2287, 2350), False, 'from dashscope import Generation\n'), ((2443, 2540), 'llama_index.legacy.bridge.pydantic.Fiel...
import os from llama_index import download_loader from llama_index.node_parser import SimpleNodeParser from llama_index import GPTVectorStoreIndex download_loader("GithubRepositoryReader") from llama_index.readers.llamahub_modules.github_repo import ( GithubRepositoryReader, GithubClient, ) # Initialize the...
[ "llama_index.GPTVectorStoreIndex", "llama_index.node_parser.SimpleNodeParser", "llama_index.download_loader", "llama_index.readers.llamahub_modules.github_repo.GithubRepositoryReader" ]
[((149, 190), 'llama_index.download_loader', 'download_loader', (['"""GithubRepositoryReader"""'], {}), "('GithubRepositoryReader')\n", (164, 190), False, 'from llama_index import download_loader\n'), ((409, 706), 'llama_index.readers.llamahub_modules.github_repo.GithubRepositoryReader', 'GithubRepositoryReader', (['gi...
"""Relevancy evaluation.""" from __future__ import annotations import asyncio from typing import Any, Optional, Sequence, Union from llama_index.core import ServiceContext from llama_index.core.evaluation.base import BaseEvaluator, EvaluationResult from llama_index.core.indices import SummaryIndex from llama_index.co...
[ "llama_index.core.prompts.PromptTemplate", "llama_index.core.indices.SummaryIndex.from_documents", "llama_index.core.evaluation.base.EvaluationResult", "llama_index.core.schema.Document", "llama_index.core.settings.llm_from_settings_or_context" ]
[((620, 974), 'llama_index.core.prompts.PromptTemplate', 'PromptTemplate', (['"""Your task is to evaluate if the response for the query is in line with the context information provided.\nYou have two options to answer. Either YES/ NO.\nAnswer - YES, if the response for the query is in line with context informat...
"""Base tool spec class.""" import asyncio from inspect import signature from typing import Any, Awaitable, Callable, Dict, List, Optional, Tuple, Type, Union from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.tools.function_tool import FunctionTool from llama_index.core.tools.types import ...
[ "llama_index.core.tools.types.ToolMetadata", "llama_index.core.tools.function_tool.FunctionTool.from_defaults" ]
[((2092, 2161), 'llama_index.core.tools.types.ToolMetadata', 'ToolMetadata', ([], {'name': 'name', 'description': 'description', 'fn_schema': 'fn_schema'}), '(name=name, description=description, fn_schema=fn_schema)\n', (2104, 2161), False, 'from llama_index.core.tools.types import ToolMetadata\n'), ((4457, 4481), 'asy...
"""Node parser interface.""" from abc import ABC, abstractmethod from typing import Any, Callable, List, Sequence from llama_index.core.bridge.pydantic import Field, validator from llama_index.core.callbacks import CallbackManager, CBEventType, EventPayload from llama_index.core.node_parser.node_utils import ( bu...
[ "llama_index.core.bridge.pydantic.Field", "llama_index.core.bridge.pydantic.validator", "llama_index.core.node_parser.node_utils.build_nodes_from_splits", "llama_index.core.utils.get_tqdm_iterable" ]
[((668, 759), 'llama_index.core.bridge.pydantic.Field', 'Field', ([], {'default': '(True)', 'description': '"""Whether or not to consider metadata when splitting."""'}), "(default=True, description=\n 'Whether or not to consider metadata when splitting.')\n", (673, 759), False, 'from llama_index.core.bridge.pydantic...
"""Tree Index inserter.""" from typing import Optional, Sequence from llama_index.core.data_structs.data_structs import IndexGraph from llama_index.core.indices.prompt_helper import PromptHelper from llama_index.core.indices.tree.utils import get_numbered_text_from_nodes from llama_index.core.indices.utils import ( ...
[ "llama_index.core.indices.prompt_helper.PromptHelper.from_llm_metadata", "llama_index.core.schema.TextNode", "llama_index.core.storage.docstore.registry.get_default_docstore", "llama_index.core.indices.tree.utils.get_numbered_text_from_nodes", "llama_index.core.indices.utils.get_sorted_node_list", "llama_...
[((5228, 5265), 'llama_index.core.indices.utils.get_sorted_node_list', 'get_sorted_node_list', (['cur_graph_nodes'], {}), '(cur_graph_nodes)\n', (5248, 5265), False, 'from llama_index.core.indices.utils import extract_numbers_given_response, get_sorted_node_list\n'), ((1733, 1788), 'llama_index.core.settings.llm_from_s...
"""JSON node parser.""" import json from typing import Any, Dict, Generator, List, Optional, Sequence from llama_index.core.callbacks.base import CallbackManager from llama_index.core.node_parser.interface import NodeParser from llama_index.core.node_parser.node_utils import build_nodes_from_splits from llama_index.co...
[ "llama_index.core.utils.get_tqdm_iterable", "llama_index.core.callbacks.base.CallbackManager" ]
[((1510, 1566), 'llama_index.core.utils.get_tqdm_iterable', 'get_tqdm_iterable', (['nodes', 'show_progress', '"""Parsing nodes"""'], {}), "(nodes, show_progress, 'Parsing nodes')\n", (1527, 1566), False, 'from llama_index.core.utils import get_tqdm_iterable\n'), ((995, 1014), 'llama_index.core.callbacks.base.CallbackMa...
import asyncio import os import tempfile import traceback from datetime import date, datetime from functools import partial from pathlib import Path import aiohttp import discord import openai import tiktoken from langchain import OpenAI from langchain.chat_models import ChatOpenAI from llama_index import ( Beauti...
[ "llama_index.SimpleDirectoryReader", "llama_index.query_engine.MultiStepQueryEngine", "llama_index.ServiceContext.from_defaults", "llama_index.OpenAIEmbedding", "llama_index.retrievers.VectorIndexRetriever", "llama_index.MockEmbedding", "llama_index.BeautifulSoupWebReader", "llama_index.QuestionAnswer...
[((1193, 1226), 'services.environment_service.EnvService.get_max_search_price', 'EnvService.get_max_search_price', ([], {}), '()\n', (1224, 1226), False, 'from services.environment_service import EnvService\n'), ((1404, 1442), 'services.environment_service.EnvService.get_google_search_api_key', 'EnvService.get_google_s...
import asyncio import json import os import tempfile import time from functools import lru_cache from logging import getLogger from pathlib import Path from fastapi import APIRouter, Request, status from fastapi.encoders import jsonable_encoder from fastapi.responses import HTMLResponse from typing import List, Dict, ...
[ "llama_index.indices.vector_store.retrievers.VectorIndexRetriever", "llama_index.vector_stores.types.MetadataInfo", "llama_index.schema.NodeWithScore", "llama_index.response_synthesizers.TreeSummarize", "llama_index.PromptTemplate", "llama_index.query_pipeline.QueryPipeline" ]
[((2834, 2845), 'logging.getLogger', 'getLogger', ([], {}), '()\n', (2843, 2845), False, 'from logging import getLogger\n'), ((2859, 2881), 'snowflake.SnowflakeGenerator', 'SnowflakeGenerator', (['(42)'], {}), '(42)\n', (2877, 2881), False, 'from snowflake import SnowflakeGenerator\n'), ((2892, 2903), 'fastapi.APIRoute...
from dotenv import load_dotenv import cv2 import numpy as np import os import streamlit as st from llama_index import SimpleDirectoryReader from pydantic_llm import ( pydantic_llm, DamagedParts, damages_initial_prompt_str, ConditionsReport, conditions_report_initial_prompt_str, ) import pandas as pd...
[ "llama_index.SimpleDirectoryReader", "llama_index.multi_modal_llms.openai.OpenAIMultiModal" ]
[((557, 607), 'streamlit_modal.Modal', 'Modal', (['"""Damage Report"""'], {'key': '"""demo"""', 'max_width': '(1280)'}), "('Damage Report', key='demo', max_width=1280)\n", (562, 607), False, 'from streamlit_modal import Modal\n'), ((912, 925), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (923, 925), False, 'f...
from typing import TYPE_CHECKING, Any, Optional from llama_index.legacy.core.base_query_engine import BaseQueryEngine if TYPE_CHECKING: from llama_index.legacy.langchain_helpers.agents.tools import ( LlamaIndexTool, ) from llama_index.legacy.tools.types import AsyncBaseTool, ToolMetadata, ToolOutput ...
[ "llama_index.legacy.langchain_helpers.agents.tools.IndexToolConfig", "llama_index.legacy.langchain_helpers.agents.tools.LlamaIndexTool.from_tool_config", "llama_index.legacy.tools.types.ToolMetadata" ]
[((1408, 1456), 'llama_index.legacy.tools.types.ToolMetadata', 'ToolMetadata', ([], {'name': 'name', 'description': 'description'}), '(name=name, description=description)\n', (1420, 1456), False, 'from llama_index.legacy.tools.types import AsyncBaseTool, ToolMetadata, ToolOutput\n'), ((3568, 3683), 'llama_index.legacy....
from llama_index.core.llama_dataset import download_llama_dataset from llama_index.core.llama_pack import download_llama_pack from llama_index.core import VectorStoreIndex async def main(): # DOWNLOAD LLAMADATASET rag_dataset, documents = download_llama_dataset( "EvaluatingLlmSurveyPaperDataset", "./d...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.core.llama_dataset.download_llama_dataset", "llama_index.core.llama_pack.download_llama_pack" ]
[((249, 316), 'llama_index.core.llama_dataset.download_llama_dataset', 'download_llama_dataset', (['"""EvaluatingLlmSurveyPaperDataset"""', '"""./data"""'], {}), "('EvaluatingLlmSurveyPaperDataset', './data')\n", (271, 316), False, 'from llama_index.core.llama_dataset import download_llama_dataset\n'), ((375, 427), 'll...
import json import os import warnings from enum import Enum from typing import Any, Callable, Dict, List, Literal, Optional, Sequence from deprecated import deprecated from llama_index.legacy.bridge.pydantic import Field, PrivateAttr from llama_index.legacy.callbacks.base import CallbackManager from llama_index.legac...
[ "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.bridge.pydantic.Field" ]
[((1210, 1271), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'description': '"""The modelId of the Bedrock model to use."""'}), "(description='The modelId of the Bedrock model to use.')\n", (1215, 1271), False, 'from llama_index.legacy.bridge.pydantic import Field, PrivateAttr\n'), ((1306, 1430), 'llama_i...
# !pip install llama-index faiss-cpu llama-index-vector-stores-faiss import faiss from llama_index.core import ( SimpleDirectoryReader, VectorStoreIndex, StorageContext, ) from llama_index.vector_stores.faiss import FaissVectorStore from llama_index.core import get_response_synthesizer from llama_index.c...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.core.query_engine.RetrieverQueryEngine", "llama_index.core.retrievers.VectorIndexRetriever", "llama_index.vector_stores.faiss.FaissVectorStore", "llama_index.core.StorageContext.from_defaults", "llama_index.core.prompts.base.PromptTemplate",...
[((1083, 1103), 'faiss.IndexFlatL2', 'faiss.IndexFlatL2', (['d'], {}), '(d)\n', (1100, 1103), False, 'import faiss\n'), ((1150, 1191), 'llama_index.vector_stores.faiss.FaissVectorStore', 'FaissVectorStore', ([], {'faiss_index': 'faiss_index'}), '(faiss_index=faiss_index)\n', (1166, 1191), False, 'from llama_index.vecto...
from dotenv import load_dotenv from llama_index.llms import OpenAI from llama_index.prompts import PromptTemplate from retriever import run_retrieval import nest_asyncio import asyncio nest_asyncio.apply() async def acombine_results( texts, query_str, qa_prompt, llm, cur_prompt_list, num_c...
[ "llama_index.prompts.PromptTemplate", "llama_index.llms.OpenAI" ]
[((189, 209), 'nest_asyncio.apply', 'nest_asyncio.apply', ([], {}), '()\n', (207, 209), False, 'import nest_asyncio\n'), ((2126, 2160), 'llama_index.llms.OpenAI', 'OpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""'}), "(model_name='gpt-3.5-turbo')\n", (2132, 2160), False, 'from llama_index.llms import OpenAI\n'), ((217...
from pathlib import Path from llama_index import download_loader from llama_index import SimpleDirectoryReader PDFReader = download_loader("PDFReader") def getdocument(filename : str,filetype:str): if filetype == "pdf": loader = PDFReader() elif filetype == "txt": loader = SimpleDirectoryReade...
[ "llama_index.SimpleDirectoryReader", "llama_index.download_loader" ]
[((124, 152), 'llama_index.download_loader', 'download_loader', (['"""PDFReader"""'], {}), "('PDFReader')\n", (139, 152), False, 'from llama_index import download_loader\n'), ((300, 334), 'llama_index.SimpleDirectoryReader', 'SimpleDirectoryReader', (['"""./example"""'], {}), "('./example')\n", (321, 334), False, 'from...
import faiss import openai from llama_index.readers.file.epub_parser import EpubParser # create an index with the text and save it to disk in data/indexes from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader, LLMPredictor from langchain.chat_models import ChatOpenAI from llama_index import GPTTreeIndex i...
[ "llama_index.GPTSimpleVectorIndex", "llama_index.SimpleDirectoryReader" ]
[((1175, 1217), 'llama_index.GPTSimpleVectorIndex', 'GPTSimpleVectorIndex', ([], {'documents': 'self._docs'}), '(documents=self._docs)\n', (1195, 1217), False, 'from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader, LLMPredictor\n'), ((1057, 1136), 'llama_index.SimpleDirectoryReader', 'SimpleDirectoryRead...
# -*- coding: utf-8 -*- # @place: Pudong, Shanghai # @file: query_rewrite_ensemble_retriever.py # @time: 2023/12/28 13:49 # -*- coding: utf-8 -*- # @place: Pudong, Shanghai # @file: ensemble_retriever.py # @time: 2023/12/26 18:50 import json from typing import List from operator import itemgetter from llama_index.sche...
[ "llama_index.schema.TextNode" ]
[((3476, 3493), 'faiss.IndexFlatIP', 'IndexFlatIP', (['(1536)'], {}), '(1536)\n', (3487, 3493), False, 'from faiss import IndexFlatIP\n'), ((3693, 3709), 'pprint.pprint', 'pprint', (['t_result'], {}), '(t_result)\n', (3699, 3709), False, 'from pprint import pprint\n'), ((942, 1030), 'custom_retriever.vector_store_retri...
"""Utils for jupyter notebook.""" import os from io import BytesIO from typing import Any, Dict, List, Tuple import matplotlib.pyplot as plt import requests from IPython.display import Markdown, display from llama_index.core.base.response.schema import Response from llama_index.core.img_utils import b64_2_img from lla...
[ "llama_index.core.img_utils.b64_2_img" ]
[((723, 741), 'llama_index.core.img_utils.b64_2_img', 'b64_2_img', (['img_str'], {}), '(img_str)\n', (732, 741), False, 'from llama_index.core.img_utils import b64_2_img\n'), ((770, 782), 'IPython.display.display', 'display', (['img'], {}), '(img)\n', (777, 782), False, 'from IPython.display import Markdown, display\n'...
from typing import Optional, Type from llama_index.legacy.download.module import ( LLAMA_HUB_URL, MODULE_TYPE, download_llama_module, track_download, ) from llama_index.legacy.llama_pack.base import BaseLlamaPack def download_llama_pack( llama_pack_class: str, download_dir: str, llama_hub...
[ "llama_index.legacy.download.module.track_download", "llama_index.legacy.download.module.download_llama_module" ]
[((887, 1134), 'llama_index.legacy.download.module.download_llama_module', 'download_llama_module', (['llama_pack_class'], {'llama_hub_url': 'llama_hub_url', 'refresh_cache': 'refresh_cache', 'custom_path': 'download_dir', 'library_path': '"""llama_packs/library.json"""', 'disable_library_cache': '(True)', 'override_pa...
# Debug stuff #import os #import readline #print("Current Working Directory:", os.getcwd()) #env_var = os.getenv('OPENAI_API_KEY') #print(env_var) # Sets llama-index import logging import sys logging.basicConfig(stream=sys.stdout, level=logging.DEBUG) logging.getLogger().addHandler(logging.StreamHandler(stream=sys.st...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.load_index_from_storage", "llama_index.SimpleDirectoryReader", "llama_index.StorageContext.from_defaults" ]
[((194, 253), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.DEBUG'}), '(stream=sys.stdout, level=logging.DEBUG)\n', (213, 253), False, 'import logging\n'), ((285, 325), 'logging.StreamHandler', 'logging.StreamHandler', ([], {'stream': 'sys.stdout'}), '(stream=sys.stdout)\...
import os, streamlit as st # Uncomment to specify your OpenAI API key here (local testing only, not in production!), or add corresponding environment variable (recommended) #os.environ['OPENAI_API_KEY']= "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" #openai_api_key = st.secrets["OPENAI_API_KEY"] from llama_i...
[ "llama_index.SimpleDirectoryReader", "llama_index.ServiceContext.from_defaults", "llama_index.GPTSimpleVectorIndex.from_documents", "llama_index.PromptHelper" ]
[((494, 519), 'streamlit.title', 'st.title', (['"""杭萧SAP交流会问答机器人"""'], {}), "('杭萧SAP交流会问答机器人')\n", (502, 519), True, 'import os, streamlit as st\n'), ((528, 595), 'streamlit.text_input', 'st.text_input', (['"""您可以询问任何关于会议内容的问题? (数据来源于两天的会议录音纪要,注:问题答案可能为英文)"""', '""""""'], {}), "('您可以询问任何关于会议内容的问题? (数据来源于两天的会议录音纪要,注:问题答...
import os, streamlit as st # Uncomment to specify your OpenAI API key here (local testing only, not in production!), or add corresponding environment variable (recommended) # os.environ['OPENAI_API_KEY']= "" from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader, LLMPredictor, PromptHelper, ServiceContex...
[ "llama_index.SimpleDirectoryReader", "llama_index.ServiceContext.from_defaults", "llama_index.GPTSimpleVectorIndex.from_documents", "llama_index.PromptHelper" ]
[((396, 417), 'streamlit.title', 'st.title', (['"""Ask Llama"""'], {}), "('Ask Llama')\n", (404, 417), True, 'import os, streamlit as st\n'), ((426, 500), 'streamlit.text_input', 'st.text_input', (['"""What would you like to ask? (source: data/Create.txt)"""', '""""""'], {}), "('What would you like to ask? (source: dat...
import os, streamlit as st # Uncomment to specify your OpenAI API key here (local testing only, not in production!), or add corresponding environment variable (recommended) # os.environ['OPENAI_API_KEY']= "" from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader, LLMPredictor, PromptHelper from langchain...
[ "llama_index.GPTSimpleVectorIndex", "llama_index.SimpleDirectoryReader", "llama_index.PromptHelper" ]
[((635, 694), 'llama_index.PromptHelper', 'PromptHelper', (['max_input_size', 'num_output', 'max_chunk_overlap'], {}), '(max_input_size, num_output, max_chunk_overlap)\n', (647, 694), False, 'from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader, LLMPredictor, PromptHelper\n'), ((801, 895), 'llama_index.G...
from typing import Any, Dict, List, Optional, Sequence, Tuple from llama_index.core.base.response.schema import RESPONSE_TYPE, Response from llama_index.core.callbacks.base import CallbackManager from llama_index.core.callbacks.schema import CBEventType, EventPayload from llama_index.core.indices.multi_modal import Mu...
[ "llama_index.core.callbacks.base.CallbackManager", "llama_index.multi_modal_llms.openai.OpenAIMultiModal" ]
[((3353, 3372), 'llama_index.core.callbacks.base.CallbackManager', 'CallbackManager', (['[]'], {}), '([])\n', (3368, 3372), False, 'from llama_index.core.callbacks.base import CallbackManager\n'), ((2707, 2774), 'llama_index.multi_modal_llms.openai.OpenAIMultiModal', 'OpenAIMultiModal', ([], {'model': '"""gpt-4-vision-...
import json from typing import Sequence from llama_index.legacy.prompts.base import PromptTemplate from llama_index.legacy.question_gen.types import SubQuestion from llama_index.legacy.tools.types import ToolMetadata # deprecated, kept for backward compatibility SubQuestionPrompt = PromptTemplate def build_tools_te...
[ "llama_index.legacy.question_gen.types.SubQuestion", "llama_index.legacy.tools.types.ToolMetadata" ]
[((465, 497), 'json.dumps', 'json.dumps', (['tools_dict'], {'indent': '(4)'}), '(tools_dict, indent=4)\n', (475, 497), False, 'import json\n'), ((817, 923), 'llama_index.legacy.tools.types.ToolMetadata', 'ToolMetadata', ([], {'name': '"""uber_10k"""', 'description': '"""Provides information about Uber financials for ye...
import os import openai from typing import Union import collections from IPython.display import Markdown, display # access/create the .env file in the project dir for getting API keys. Create a .env file in the project/repository root, # and add your own API key like "OPENAI_API_KEY = <your key>" without any quotes, ...
[ "llama_index.objects.SQLTableNodeMapping", "llama_index.ServiceContext.from_defaults", "llama_index.logger.LlamaLogger", "llama_index.objects.SQLTableSchema", "llama_index.set_global_service_context", "llama_index.SQLDatabase", "llama_index.objects.ObjectIndex.from_objects" ]
[((517, 530), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (528, 530), False, 'from dotenv import load_dotenv\n'), ((2230, 2243), 'llama_index.logger.LlamaLogger', 'LlamaLogger', ([], {}), '()\n', (2241, 2243), False, 'from llama_index.logger import LlamaLogger\n'), ((2277, 2304), 'os.getenv', 'os.getenv', ([...
# # Copyright DataStax, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, soft...
[ "llama_index.VectorStoreIndex.from_vector_store", "llama_index.vector_stores.CassandraVectorStore" ]
[((1311, 1342), 'base64.b64decode', 'base64.b64decode', (['secure_bundle'], {}), '(secure_bundle)\n', (1327, 1342), False, 'import base64\n'), ((1765, 1955), 'llama_index.vector_stores.CassandraVectorStore', 'CassandraVectorStore', ([], {'session': 'self.session', 'keyspace': "self.config['cassandra']['keyspace']", 'ta...
import os from django.conf import settings from postdata.models import UploadedFile from .create_node import * import llama_index from llama_index.llms import OpenAI from llama_index import (VectorStoreIndex, ServiceContext, set_global_service_context, ...
[ "llama_index.ServiceContext.from_defaults", "llama_index.set_global_service_context", "llama_index.set_global_handler", "llama_index.VectorStoreIndex" ]
[((334, 374), 'llama_index.set_global_handler', 'llama_index.set_global_handler', (['"""simple"""'], {}), "('simple')\n", (364, 374), False, 'import llama_index\n'), ((544, 581), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'llm': 'llm'}), '(llm=llm)\n', (572, 581), False, 'from lla...
from typing import Any, List, Optional from llama_index.legacy.bridge.pydantic import Field, PrivateAttr from llama_index.legacy.callbacks import CallbackManager from llama_index.legacy.core.embeddings.base import ( DEFAULT_EMBED_BATCH_SIZE, BaseEmbedding, ) from llama_index.legacy.embeddings.huggingface_utils...
[ "llama_index.legacy.embeddings.huggingface_utils.get_text_instruct_for_model_name", "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.bridge.pydantic.Field", "llama_index.legacy.embeddings.huggingface_utils.get_query_instruct_for_model_name" ]
[((520, 578), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'description': '"""Instruction to prepend to query text."""'}), "(description='Instruction to prepend to query text.')\n", (525, 578), False, 'from llama_index.legacy.bridge.pydantic import Field, PrivateAttr\n'), ((631, 683), 'llama_index.legacy....
"""Base retrieval abstractions.""" import asyncio from abc import abstractmethod from enum import Enum from typing import Any, Dict, List, Optional, Tuple from llama_index.core.bridge.pydantic import BaseModel, Field from llama_index.core.evaluation.retrieval.metrics import resolve_metrics from llama_index.core.evalu...
[ "llama_index.core.bridge.pydantic.Field", "llama_index.core.evaluation.retrieval.metrics.resolve_metrics" ]
[((1364, 1402), 'llama_index.core.bridge.pydantic.Field', 'Field', (['...'], {'description': '"""Query string"""'}), "(..., description='Query string')\n", (1369, 1402), False, 'from llama_index.core.bridge.pydantic import BaseModel, Field\n'), ((1433, 1471), 'llama_index.core.bridge.pydantic.Field', 'Field', (['...'],...
"""Code splitter.""" from typing import Any, Callable, List, Optional from llama_index.legacy.bridge.pydantic import Field, PrivateAttr from llama_index.legacy.callbacks.base import CallbackManager from llama_index.legacy.callbacks.schema import CBEventType, EventPayload from llama_index.legacy.node_parser.interface ...
[ "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.bridge.pydantic.Field", "llama_index.legacy.callbacks.base.CallbackManager" ]
[((779, 849), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'description': '"""The programming language of the code being split."""'}), "(description='The programming language of the code being split.')\n", (784, 849), False, 'from llama_index.legacy.bridge.pydantic import Field, PrivateAttr\n'), ((887, 99...
import asyncio from llama_index.core.llama_dataset import download_llama_dataset from llama_index.core.llama_pack import download_llama_pack from llama_index.core import VectorStoreIndex from llama_index.llms import OpenAI async def main(): # DOWNLOAD LLAMADATASET rag_dataset, documents = download_llama_data...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.llms.OpenAI", "llama_index.core.llama_dataset.download_llama_dataset", "llama_index.core.llama_pack.download_llama_pack" ]
[((301, 376), 'llama_index.core.llama_dataset.download_llama_dataset', 'download_llama_dataset', (['"""DocugamiKgRagSec10Q"""', '"""./docugami_kg_rag_sec_10_q"""'], {}), "('DocugamiKgRagSec10Q', './docugami_kg_rag_sec_10_q')\n", (323, 376), False, 'from llama_index.core.llama_dataset import download_llama_dataset\n'), ...
import os import torch import json import argparse from datasets import load_dataset from llama_index import GPTVectorStoreIndex, Document, ServiceContext from llama_index.indices.prompt_helper import PromptHelper from transformers import AutoTokenizer import openai import tiktoken #import GPUtil stopped_num = 10000000...
[ "llama_index.indices.prompt_helper.PromptHelper", "llama_index.ServiceContext.from_defaults", "llama_index.GPTVectorStoreIndex.from_documents", "llama_index.Document" ]
[((783, 808), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (806, 808), False, 'import argparse\n'), ((1533, 1569), 'tiktoken.get_encoding', 'tiktoken.get_encoding', (['encoding_name'], {}), '(encoding_name)\n', (1554, 1569), False, 'import tiktoken\n'), ((1870, 1988), 'llama_index.indices.pro...
# inspired by: https://github.com/rushic24/langchain-remember-me-llm/ # MIT license import torch from json_database import JsonStorageXDG from langchain.embeddings.huggingface import HuggingFaceEmbeddings from langchain.llms.base import LLM from llama_index import Document from llama_index import LLMPredictor, ServiceC...
[ "llama_index.ServiceContext.from_defaults", "llama_index.GPTVectorStoreIndex.from_documents", "llama_index.Document", "llama_index.LangchainEmbedding" ]
[((541, 570), 'json_database.JsonStorageXDG', 'JsonStorageXDG', (['"""personalLLM"""'], {}), "('personalLLM')\n", (555, 570), False, 'from json_database import JsonStorageXDG\n'), ((1152, 1263), 'transformers.pipeline', 'pipeline', (['"""text2text-generation"""'], {'model': 'model_name', 'device': '(0)', 'model_kwargs'...
from dotenv import load_dotenv import os.path from llama_index.core import ( VectorStoreIndex, SimpleDirectoryReader, StorageContext, load_index_from_storage, ) import logging import sys load_dotenv() logging.basicConfig(stream=sys.stdout, level=logging.INFO) logging.getLogger().addHandler(logging.Str...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.core.StorageContext.from_defaults", "llama_index.core.load_index_from_storage", "llama_index.core.SimpleDirectoryReader" ]
[((204, 217), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (215, 217), False, 'from dotenv import load_dotenv\n'), ((219, 277), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.INFO'}), '(stream=sys.stdout, level=logging.INFO)\n', (238, 277), False, 'import logging...
from typing import Literal from llama_index.core.schema import BaseNode, TextNode from transformers import AutoModelForSeq2SeqLM, AutoTokenizer from llama_index.core.extractors import TitleExtractor, QuestionsAnsweredExtractor from llama_index.core import Document, Node from llama_index.core.node_parser.file.html impor...
[ "llama_index.core.node_parser.file.html.DEFAULT_TAGS.copy", "llama_index.core.node_parser.file.html.HTMLNodeParser", "llama_index.core.schema.TextNode", "llama_index.extractors.entity.EntityExtractor" ]
[((6318, 6362), 'src.file_utils.createOutputFile', 'createOutputFile', (['"""./kg-output"""', '"""token-gen"""'], {}), "('./kg-output', 'token-gen')\n", (6334, 6362), False, 'from src.file_utils import createOutputFile\n'), ((14566, 14643), 'llama_index.extractors.entity.EntityExtractor', 'EntityExtractor', ([], {'pred...
"""Table node mapping.""" from typing import Any, Dict, Optional, Sequence from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.objects.base_node_mapping import ( DEFAULT_PERSIST_DIR, DEFAULT_PERSIST_FNAME, BaseObjectNodeMapping, ) from llama_index.core.schema import BaseNode, Text...
[ "llama_index.core.schema.TextNode" ]
[((1821, 1968), 'llama_index.core.schema.TextNode', 'TextNode', ([], {'text': 'table_text', 'metadata': 'metadata', 'excluded_embed_metadata_keys': "['name', 'context']", 'excluded_llm_metadata_keys': "['name', 'context']"}), "(text=table_text, metadata=metadata, excluded_embed_metadata_keys=[\n 'name', 'context'], ...
"""Base query engine.""" import logging from abc import abstractmethod from typing import Any, Dict, List, Optional, Sequence from llama_index.legacy.bridge.pydantic import Field from llama_index.legacy.callbacks.base import CallbackManager from llama_index.legacy.core.query_pipeline.query_component import ( Chai...
[ "llama_index.legacy.core.query_pipeline.query_component.validate_and_convert_stringable", "llama_index.legacy.core.query_pipeline.query_component.InputKeys.from_keys", "llama_index.legacy.schema.QueryBundle", "llama_index.legacy.core.query_pipeline.query_component.OutputKeys.from_keys", "llama_index.legacy....
[((647, 674), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (664, 674), False, 'import logging\n'), ((3066, 3104), 'llama_index.legacy.bridge.pydantic.Field', 'Field', (['...'], {'description': '"""Query engine"""'}), "(..., description='Query engine')\n", (3071, 3104), False, 'from llam...
import os from shutil import rmtree from typing import Callable, Dict, List, Optional import tqdm from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.postprocessor.types import BaseNodePostprocessor from llama_index.core.schema import Document, QueryBundle from llama_index.core.utils i...
[ "llama_index.core.schema.Document", "llama_index.core.utils.get_cache_dir", "llama_index.core.schema.QueryBundle" ]
[((861, 876), 'llama_index.core.utils.get_cache_dir', 'get_cache_dir', ([], {}), '()\n', (874, 876), False, 'from llama_index.core.utils import get_cache_dir\n'), ((970, 1025), 'os.path.join', 'os.path.join', (['cache_dir', '"""datasets"""', "('BeIR__' + dataset)"], {}), "(cache_dir, 'datasets', 'BeIR__' + dataset)\n",...
import logging from typing import Any, Dict, Generator, List, Optional, Tuple, Type, Union, cast from llama_index.legacy.agent.openai.utils import resolve_tool_choice from llama_index.legacy.llms.llm import LLM from llama_index.legacy.llms.openai import OpenAI from llama_index.legacy.llms.openai_utils import OpenAIToo...
[ "llama_index.legacy.program.utils.create_list_model", "llama_index.legacy.agent.openai.utils.resolve_tool_choice", "llama_index.legacy.llms.openai.OpenAI", "llama_index.legacy.llms.openai_utils.to_openai_tool", "llama_index.legacy.prompts.base.PromptTemplate" ]
[((619, 646), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (636, 646), False, 'import logging\n'), ((914, 950), 'llama_index.legacy.agent.openai.utils.resolve_tool_choice', 'resolve_tool_choice', (["schema['title']"], {}), "(schema['title'])\n", (933, 950), False, 'from llama_index.lega...
# use SQLAlchemy to setup a simple sqlite db from sqlalchemy import (Column, Integer, MetaData, String, Table, column, create_engine, select) engine = create_engine("sqlite:///:memory:") metadata_obj = MetaData() # create a toy city_stats table table_name = "city_stats" city_stats_table = Tabl...
[ "llama_index.SQLDatabase" ]
[((176, 211), 'sqlalchemy.create_engine', 'create_engine', (['"""sqlite:///:memory:"""'], {}), "('sqlite:///:memory:')\n", (189, 211), False, 'from sqlalchemy import Column, Integer, MetaData, String, Table, column, create_engine, select\n'), ((227, 237), 'sqlalchemy.MetaData', 'MetaData', ([], {}), '()\n', (235, 237),...
from llama_index import SimpleDirectoryReader, GPTSimpleVectorIndex, LLMPredictor, PromptHelper from langchain import OpenAI class GPTModel: def __init__(self, directory_path): # set maximum input size self.max_input_size = 4096 # set number of output tokens self.num_outputs = 2000 ...
[ "llama_index.GPTSimpleVectorIndex", "llama_index.SimpleDirectoryReader", "llama_index.PromptHelper" ]
[((673, 792), 'llama_index.PromptHelper', 'PromptHelper', (['self.max_input_size', 'self.num_outputs', 'self.max_chunk_overlap'], {'chunk_size_limit': 'self.chunk_size_limit'}), '(self.max_input_size, self.num_outputs, self.max_chunk_overlap,\n chunk_size_limit=self.chunk_size_limit)\n', (685, 792), False, 'from lla...
from typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, cast import httpx from openai import AsyncOpenAI from openai import OpenAI as SyncOpenAI from openai.types.chat import ChatCompletionMessageParam from openai.types.chat.chat_completion_chunk import ( ChatCompletionChunk, ChoiceDelta, ...
[ "llama_index.legacy.core.llms.types.ChatMessage", "llama_index.legacy.multi_modal_llms.openai_utils.GPT4V_MODELS.keys", "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.multi_modal_llms.openai_utils.generate_openai_multi_modal_chat_message", "llama_index.legacy.llms.openai_utils.from_op...
[((1407, 1469), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'description': '"""The Multi-Modal model to use from OpenAI."""'}), "(description='The Multi-Modal model to use from OpenAI.')\n", (1412, 1469), False, 'from llama_index.legacy.bridge.pydantic import Field, PrivateAttr\n'), ((1495, 1552), 'llama...
import os from typing import Any from llama_index import ServiceContext, VectorStoreIndex from llama_index.embeddings.openai import OpenAIEmbedding, OpenAIEmbeddingMode from llama_index.prompts import PromptTemplate from llama_index.indices.query.schema import QueryBundle from llama_index.llms import OpenAI from llama...
[ "llama_index.prompts.PromptTemplate", "llama_index.llms.OpenAI", "llama_index.VectorStoreIndex", "llama_index.VectorStoreIndex.from_vector_store", "llama_index.embeddings.openai.OpenAIEmbedding" ]
[((2518, 2654), 'llama_index.VectorStoreIndex.from_vector_store', 'VectorStoreIndex.from_vector_store', (['docstore.vector_store'], {'service_context': 'self.service_context', 'show_progress': '(True)', 'use_async': '(True)'}), '(docstore.vector_store, service_context=\n self.service_context, show_progress=True, use...
import os from dotenv import load_dotenv, find_dotenv import numpy as np from trulens_eval import ( Feedback, TruLlama, OpenAI ) from trulens_eval.feedback import Groundedness import nest_asyncio nest_asyncio.apply() def get_openai_api_key(): _ = load_dotenv(find_dotenv()) return os.getenv("OP...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.retrievers.AutoMergingRetriever", "llama_index.node_parser.HierarchicalNodeParser.from_defaults", "llama_index.VectorStoreIndex", "llama_index.indices.postprocessor.SentenceTransformerRerank", "llama_index.node_parser.SentenceWindowNodeParser.fro...
[((211, 231), 'nest_asyncio.apply', 'nest_asyncio.apply', ([], {}), '()\n', (229, 231), False, 'import nest_asyncio\n'), ((449, 457), 'trulens_eval.OpenAI', 'OpenAI', ([], {}), '()\n', (455, 457), False, 'from trulens_eval import Feedback, TruLlama, OpenAI\n'), ((854, 896), 'trulens_eval.feedback.Groundedness', 'Ground...
"""SQL Structured Store.""" from collections import defaultdict from enum import Enum from typing import Any, Optional, Sequence, Union from sqlalchemy import Table from llama_index.legacy.core.base_query_engine import BaseQueryEngine from llama_index.legacy.core.base_retriever import BaseRetriever from llama_index....
[ "llama_index.legacy.indices.struct_store.container_builder.SQLContextContainerBuilder", "llama_index.legacy.indices.struct_store.sql_query.NLStructStoreQueryEngine", "llama_index.legacy.indices.struct_store.sql_query.SQLStructStoreQueryEngine", "llama_index.legacy.indices.common.struct_store.sql.SQLStructData...
[((5106, 5332), 'llama_index.legacy.indices.common.struct_store.sql.SQLStructDatapointExtractor', 'SQLStructDatapointExtractor', (['self._service_context.llm', 'self.schema_extract_prompt', 'self.output_parser', 'self.sql_database'], {'table_name': 'self._table_name', 'table': 'self._table', 'ref_doc_id_column': 'self....
from llama_index.indices.multi_modal.base import MultiModalVectorStoreIndex from llama_index import SimpleDirectoryReader, StorageContext from usearch.index import Index from fast_mm_rag import ClipCppEmbedding, USearchVectorStore from PIL import Image import matplotlib.pyplot as plt import os def plot_images(image...
[ "llama_index.SimpleDirectoryReader", "llama_index.response.notebook_utils.display_source_node", "llama_index.StorageContext.from_defaults", "llama_index.indices.multi_modal.base.MultiModalVectorStoreIndex.from_documents" ]
[((731, 760), 'usearch.index.Index', 'Index', ([], {'ndim': '(512)', 'metric': '"""cos"""'}), "(ndim=512, metric='cos')\n", (736, 760), False, 'from usearch.index import Index\n'), ((774, 821), 'fast_mm_rag.USearchVectorStore', 'USearchVectorStore', ([], {'usearch_index': 'usearch_index'}), '(usearch_index=usearch_inde...
"""Base vector store index query.""" from pathlib import Path from typing import List, Optional from llama_index import QueryBundle, StorageContext, load_index_from_storage from llama_index.data_structs import NodeWithScore, IndexDict from llama_index.indices.utils import log_vector_store_query_result from llama_index...
[ "llama_index.vector_stores.FaissVectorStore.from_persist_dir", "llama_index.StorageContext.from_defaults", "llama_index.data_structs.NodeWithScore", "llama_index.vector_stores.types.VectorStoreQuery", "llama_index.indices.utils.log_vector_store_query_result", "llama_index.load_index_from_storage", "llam...
[((1342, 1371), 'llama_index.token_counter.token_counter.llm_token_counter', 'llm_token_counter', (['"""retrieve"""'], {}), "('retrieve')\n", (1359, 1371), False, 'from llama_index.token_counter.token_counter import llm_token_counter\n'), ((1813, 2059), 'llama_index.vector_stores.types.VectorStoreQuery', 'VectorStoreQu...
import logging import sys import os.path from llama_index.core import ( VectorStoreIndex, SimpleDirectoryReader, StorageContext, load_index_from_storage, ) logging.basicConfig(stream=sys.stdout, level=logging.DEBUG) logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stdout)) # check if s...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.core.StorageContext.from_defaults", "llama_index.core.load_index_from_storage", "llama_index.core.SimpleDirectoryReader" ]
[((173, 232), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.DEBUG'}), '(stream=sys.stdout, level=logging.DEBUG)\n', (192, 232), False, 'import logging\n'), ((264, 304), 'logging.StreamHandler', 'logging.StreamHandler', ([], {'stream': 'sys.stdout'}), '(stream=sys.stdout)\...
import os from llama_index import GPTVectorStoreIndex, SimpleDirectoryReader from IPython.display import Markdown, display from llama_index import StorageContext, load_index_from_storage # Set the OPENAI_API_KEY environment variable using the value from st.secrets['OPENAI_API_KEY'] os.environ['OPENAI_API_KEY'] = st.se...
[ "llama_index.SimpleDirectoryReader", "llama_index.GPTVectorStoreIndex.from_documents" ]
[((495, 540), 'llama_index.GPTVectorStoreIndex.from_documents', 'GPTVectorStoreIndex.from_documents', (['documents'], {}), '(documents)\n', (529, 540), False, 'from llama_index import GPTVectorStoreIndex, SimpleDirectoryReader\n'), ((400, 429), 'llama_index.SimpleDirectoryReader', 'SimpleDirectoryReader', (['"""data"""...
from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext from llama_index.embeddings import resolve_embed_model # Don't Import "from openai import OpenAI". It will panic from llama_index.llms import OpenAI # load data documents = SimpleDirectoryReader("data").load_data() # bge-m3 embedding mod...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.SimpleDirectoryReader", "llama_index.ServiceContext.from_defaults", "llama_index.llms.OpenAI", "llama_index.embeddings.resolve_embed_model" ]
[((337, 388), 'llama_index.embeddings.resolve_embed_model', 'resolve_embed_model', (['"""local:BAAI/bge-small-en-v1.5"""'], {}), "('local:BAAI/bge-small-en-v1.5')\n", (356, 388), False, 'from llama_index.embeddings import resolve_embed_model\n'), ((412, 477), 'llama_index.llms.OpenAI', 'OpenAI', ([], {'api_base': '"""h...
import os import openai from dotenv import load_dotenv from llama_index.embeddings import AzureOpenAIEmbedding, OpenAIEmbedding from llama_index.llms import AzureOpenAI, OpenAI, OpenAILike from llama_index.llms.llama_utils import messages_to_prompt def load_models(args, logger): llm_service = args.llm_service ...
[ "llama_index.embeddings.AzureOpenAIEmbedding", "llama_index.llms.OpenAI", "llama_index.llms.AzureOpenAI", "llama_index.embeddings.OpenAIEmbedding", "llama_index.llms.OpenAILike" ]
[((353, 366), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (364, 366), False, 'from dotenv import load_dotenv\n'), ((550, 584), 'os.getenv', 'os.getenv', (['"""AZURE_OPENAI_GPT4_KEY"""'], {}), "('AZURE_OPENAI_GPT4_KEY')\n", (559, 584), False, 'import os\n'), ((1760, 1777), 'llama_index.embeddings.OpenAIEmbedd...
from llama_index.core.tools import FunctionTool import os note_file = os.path.join("data", "notes.txt") def save_note(note): if not os.path.exists(note_file): open(note_file, "w") with open(note_file, "a") as f: f.writelines([note + "\n"]) return "note saved" note_engine = FunctionToo...
[ "llama_index.core.tools.FunctionTool.from_defaults" ]
[((71, 104), 'os.path.join', 'os.path.join', (['"""data"""', '"""notes.txt"""'], {}), "('data', 'notes.txt')\n", (83, 104), False, 'import os\n'), ((309, 448), 'llama_index.core.tools.FunctionTool.from_defaults', 'FunctionTool.from_defaults', ([], {'fn': 'save_note', 'name': '"""note_saver"""', 'description': '"""this ...
from llama_index import VectorStoreIndex, download_loader, StorageContext from llama_index.vector_stores import PineconeVectorStore """Simple reader that reads wikipedia.""" from typing import Any, List from llama_index.readers.base import BaseReader from llama_index.schema import Document from dotenv import load_do...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.schema.Document", "llama_index.download_loader" ]
[((366, 379), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (377, 379), False, 'from dotenv import load_dotenv\n'), ((1291, 1325), 'llama_index.download_loader', 'download_loader', (['"""WikipediaReader"""'], {}), "('WikipediaReader')\n", (1306, 1325), False, 'from llama_index import VectorStoreIndex, download...
# Load indices from disk from llama_index.core import load_index_from_storage from llama_index.core import StorageContext from llama_index.core.tools import QueryEngineTool, ToolMetadata from llama_index.llms.openai import OpenAI from llama_index.core.query_engine import SubQuestionQueryEngine from llama_index.agent.op...
[ "llama_index.core.tools.ToolMetadata", "llama_index.agent.openai.OpenAIAgent.from_tools", "llama_index.core.load_index_from_storage", "llama_index.llms.openai.OpenAI" ]
[((452, 491), 'os.path.join', 'os.path.join', (['script_dir', '"""config.json"""'], {}), "(script_dir, 'config.json')\n", (464, 491), False, 'import os\n'), ((562, 609), 'os.path.join', 'os.path.join', (['script_dir', "config['storage-dir']"], {}), "(script_dir, config['storage-dir'])\n", (574, 609), False, 'import os\...
import logging logging.basicConfig(level=logging.CRITICAL) import os from pathlib import Path import openai from dotenv import load_dotenv from langchain.chat_models import ChatOpenAI from llama_index import ( GPTVectorStoreIndex, LLMPredictor, ServiceContext, StorageContext, download_loader, ...
[ "llama_index.ServiceContext.from_defaults", "llama_index.GPTVectorStoreIndex.from_documents", "llama_index.load_index_from_storage", "llama_index.download_loader" ]
[((16, 59), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.CRITICAL'}), '(level=logging.CRITICAL)\n', (35, 59), False, 'import logging\n'), ((444, 457), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (455, 457), False, 'from dotenv import load_dotenv\n'), ((644, 729), 'llama_index.Service...
from llama_index.core.node_parser import SentenceWindowNodeParser from llama_index.readers.file import FlatReader from pathlib import Path reader = FlatReader() document = reader.load_data(Path("files/sample_document1.txt")) parser = SentenceWindowNodeParser.from_defaults( window_size=2, window_metadata_key...
[ "llama_index.core.node_parser.SentenceWindowNodeParser.from_defaults", "llama_index.readers.file.FlatReader" ]
[((149, 161), 'llama_index.readers.file.FlatReader', 'FlatReader', ([], {}), '()\n', (159, 161), False, 'from llama_index.readers.file import FlatReader\n'), ((236, 377), 'llama_index.core.node_parser.SentenceWindowNodeParser.from_defaults', 'SentenceWindowNodeParser.from_defaults', ([], {'window_size': '(2)', 'window_...
# uses brave (requires api key) for web search then uses ollama for local embedding and inference, for a cost-free web RAG # requires ollama to be installed and running import os import json import logging import sys import requests from dotenv import load_dotenv from requests.adapters import HTTPAdapter from urllib3....
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.llms.ollama.Ollama", "llama_index.tools.brave_search.BraveSearchToolSpec", "llama_index.core.Document", "llama_index.embeddings.ollama.OllamaEmbedding" ]
[((660, 706), 'llama_index.embeddings.ollama.OllamaEmbedding', 'OllamaEmbedding', ([], {'model_name': '"""nomic-embed-text"""'}), "(model_name='nomic-embed-text')\n", (675, 706), False, 'from llama_index.embeddings.ollama import OllamaEmbedding\n'), ((814, 860), 'llama_index.llms.ollama.Ollama', 'Ollama', ([], {'model'...
import os os.environ["HF_HOME"] = os.path.join(os.getcwd(), "huggingface_cache") os.environ["HF_TOKEN"] = "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" os.environ["LANCEDB_CONFIG_DIR"] = os.path.join(os.getcwd(), "lancedb_config") os.environ["PYTORCH_KERNEL_CACHE_PATH"] = os.path.join(os.getcwd(), "pytorch_kernel_cache") i...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.embeddings.huggingface.HuggingFaceEmbedding", "llama_index.core.StorageContext.from_defaults", "llama_index.vector_stores.lancedb.LanceDBVectorStore", "llama_index.llms.llama_cpp.LlamaCPP", "llama_index.core.SimpleDirectoryReader" ]
[((48, 59), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (57, 59), False, 'import os\n'), ((195, 206), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (204, 206), False, 'import os\n'), ((281, 292), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (290, 292), False, 'import os\n'), ((946, 1213), 'llama_index.llms.llama_cpp.Llama...
from llama_index import ( load_index_from_storage, ServiceContext, StorageContext, LangchainEmbedding, ) from llama_index.tools import QueryEngineTool, ToolMetadata from llama_index.query_engine import SubQuestionQueryEngine import os from langchain.embeddings.huggingface import HuggingFaceEmbedding...
[ "llama_index.ServiceContext.from_defaults", "llama_index.tools.ToolMetadata", "llama_index.StorageContext.from_defaults", "llama_index.load_index_from_storage", "llama_index.query_engine.SubQuestionQueryEngine.from_defaults" ]
[((983, 1042), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'embed_model': 'query_embed_model'}), '(embed_model=query_embed_model)\n', (1011, 1042), False, 'from llama_index import load_index_from_storage, ServiceContext, StorageContext, LangchainEmbedding\n'), ((1130, 1193), 'llama...
import tkinter as tk from tkinter import filedialog from llama_index import GPTSimpleVectorIndex, SimpleDirectoryReader import os os.environ['OPENAI_API_KEY'] = 'sk-'# Your API key class MyApp(tk.Frame): def __init__(self, master=None): super().__init__(master) self.master = master ...
[ "llama_index.GPTSimpleVectorIndex", "llama_index.SimpleDirectoryReader", "llama_index.GPTSimpleVectorIndex.load_from_disk" ]
[((3123, 3130), 'tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (3128, 3130), True, 'import tkinter as tk\n'), ((505, 591), 'tkinter.Label', 'tk.Label', (['self'], {'text': '"""Document Chatbot"""', 'font': "('Arial', 16, 'bold')", 'bg': '"""#f0f0f0"""'}), "(self, text='Document Chatbot', font=('Arial', 16, 'bold'), bg=\n ...
"""Composability graphs.""" from typing import Any, Dict, List, Optional, Sequence, Type, cast from llama_index.legacy.core.base_query_engine import BaseQueryEngine from llama_index.legacy.data_structs.data_structs import IndexStruct from llama_index.legacy.indices.base import BaseIndex from llama_index.legacy.schema...
[ "llama_index.legacy.query_engine.graph_query_engine.ComposableGraphQueryEngine", "llama_index.legacy.schema.RelatedNodeInfo", "llama_index.legacy.service_context.ServiceContext.from_defaults" ]
[((4914, 4956), 'llama_index.legacy.query_engine.graph_query_engine.ComposableGraphQueryEngine', 'ComposableGraphQueryEngine', (['self'], {}), '(self, **kwargs)\n', (4940, 4956), False, 'from llama_index.legacy.query_engine.graph_query_engine import ComposableGraphQueryEngine\n'), ((1930, 1960), 'llama_index.legacy.ser...
from langchain.callbacks import CallbackManager from llama_index import ServiceContext, PromptHelper, LLMPredictor from core.callback_handler.std_out_callback_handler import DifyStdOutCallbackHandler from core.embedding.openai_embedding import OpenAIEmbedding from core.llm.llm_builder import LLMBuilder class IndexBui...
[ "llama_index.PromptHelper", "llama_index.LLMPredictor" ]
[((599, 745), 'core.llm.llm_builder.LLMBuilder.to_llm', 'LLMBuilder.to_llm', ([], {'tenant_id': 'tenant_id', 'model_name': '"""text-davinci-003"""', 'temperature': '(0)', 'max_tokens': 'num_output', 'callback_manager': 'callback_manager'}), "(tenant_id=tenant_id, model_name='text-davinci-003',\n temperature=0, max_t...
#main.py from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext from llama_index.embeddings import resolve_embed_model from llama_index.llms import OpenAI documents = SimpleDirectoryReader("data-qas").load_data() embed_model = resolve_embed_model("local:BAAI/bge-small-en-v1.5") llm = OpenAI(...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.SimpleDirectoryReader", "llama_index.ServiceContext.from_defaults", "llama_index.llms.OpenAI", "llama_index.embeddings.resolve_embed_model" ]
[((254, 305), 'llama_index.embeddings.resolve_embed_model', 'resolve_embed_model', (['"""local:BAAI/bge-small-en-v1.5"""'], {}), "('local:BAAI/bge-small-en-v1.5')\n", (273, 305), False, 'from llama_index.embeddings import resolve_embed_model\n'), ((313, 400), 'llama_index.llms.OpenAI', 'OpenAI', ([], {'temperature': '(...
import os from llama_index import VectorStoreIndex, StorageContext, \ load_indices_from_storage, ServiceContext from common.config import index_dir from common.llm import create_llm from common.utils import find_typed title = "北京市" storage_context = StorageContext.from_defaults(persist_dir=os.path.join(index_dir...
[ "llama_index.load_indices_from_storage" ]
[((405, 501), 'llama_index.load_indices_from_storage', 'load_indices_from_storage', ([], {'storage_context': 'storage_context', 'service_context': 'service_context'}), '(storage_context=storage_context, service_context=\n service_context)\n', (430, 501), False, 'from llama_index import VectorStoreIndex, StorageConte...
"""Default prompt selectors.""" from llama_index.core.prompts import SelectorPromptTemplate from llama_index.core.prompts.chat_prompts import ( CHAT_REFINE_PROMPT, CHAT_REFINE_TABLE_CONTEXT_PROMPT, CHAT_TEXT_QA_PROMPT, CHAT_TREE_SUMMARIZE_PROMPT, ) from llama_index.core.prompts.default_prompts import ( ...
[ "llama_index.core.prompts.SelectorPromptTemplate" ]
[((540, 660), 'llama_index.core.prompts.SelectorPromptTemplate', 'SelectorPromptTemplate', ([], {'default_template': 'DEFAULT_TEXT_QA_PROMPT', 'conditionals': '[(is_chat_model, CHAT_TEXT_QA_PROMPT)]'}), '(default_template=DEFAULT_TEXT_QA_PROMPT,\n conditionals=[(is_chat_model, CHAT_TEXT_QA_PROMPT)])\n', (562, 660), ...
"""Langchain memory wrapper (for LlamaIndex).""" from typing import Any, Dict, List, Optional from llama_index.core.bridge.langchain import ( AIMessage, BaseChatMemory, BaseMessage, HumanMessage, ) from llama_index.core.bridge.langchain import BaseMemory as Memory from llama_index.core.bridge.pydantic...
[ "llama_index.core.bridge.pydantic.Field", "llama_index.core.bridge.langchain.HumanMessage", "llama_index.core.bridge.langchain.AIMessage", "llama_index.core.schema.Document" ]
[((1663, 1690), 'llama_index.core.bridge.pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1668, 1690), False, 'from llama_index.core.bridge.pydantic import Field\n'), ((4306, 4333), 'llama_index.core.bridge.pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_f...
import matplotlib.pyplot as plt import polars as pl import seaborn as sns import torch from llama_index.evaluation import RelevancyEvaluator from llama_index.llms import HuggingFaceLLM from llama_index.prompts import PromptTemplate from tqdm import tqdm from transformers import BitsAndBytesConfig from src.common.utils...
[ "llama_index.prompts.PromptTemplate", "llama_index.evaluation.RelevancyEvaluator" ]
[((376, 415), 'polars.Config.set_tbl_formatting', 'pl.Config.set_tbl_formatting', (['"""NOTHING"""'], {}), "('NOTHING')\n", (404, 415), True, 'import polars as pl\n'), ((416, 441), 'polars.Config.set_tbl_rows', 'pl.Config.set_tbl_rows', (['(4)'], {}), '(4)\n', (438, 441), True, 'import polars as pl\n'), ((535, 579), 's...
import uuid from llama_index import (StorageContext, VectorStoreIndex, download_loader, load_index_from_storage) from llama_index.memory import ChatMemoryBuffer def create_index_and_query(transcript_id: str, full_transcription: any): persist_dir = f'./storage/cache/transcription/{transcr...
[ "llama_index.memory.ChatMemoryBuffer.from_defaults", "llama_index.VectorStoreIndex.from_documents", "llama_index.download_loader", "llama_index.StorageContext.from_defaults", "llama_index.load_index_from_storage" ]
[((963, 1011), 'llama_index.memory.ChatMemoryBuffer.from_defaults', 'ChatMemoryBuffer.from_defaults', ([], {'token_limit': '(2000)'}), '(token_limit=2000)\n', (993, 1011), False, 'from llama_index.memory import ChatMemoryBuffer\n'), ((365, 418), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults'...
from dotenv import load_dotenv import os # for env variables import logging import sys logging.basicConfig(stream=sys.stdout, level=logging.INFO) logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stdout)) # Load environment variables load_dotenv() PINECONE_API_KEY = os.getenv('PINECONE_API_KEY') PINECONE...
[ "llama_index.llms.ollama.Ollama", "llama_index.core.VectorStoreIndex.from_vector_store", "llama_index.vector_stores.pinecone.PineconeVectorStore", "llama_index.core.query_pipeline.InputComponent", "llama_index.core.response_synthesizers.TreeSummarize", "llama_index.core.query_pipeline.QueryPipeline", "l...
[((87, 145), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.INFO'}), '(stream=sys.stdout, level=logging.INFO)\n', (106, 145), False, 'import logging\n'), ((249, 262), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (260, 262), False, 'from dotenv import load_dotenv\...
import glob import os import re from PIL import Image from io import BytesIO from openai import OpenAI from llama_index.node_parser import MarkdownNodeParser from llama_index import ServiceContext, VectorStoreIndex, SimpleDirectoryReader from llama_index.embeddings import OpenAIEmbedding from llama_index import downloa...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.SimpleDirectoryReader", "llama_index.download_loader", "llama_index.ServiceContext.from_defaults", "llama_index.embeddings.OpenAIEmbedding", "llama_index.node_parser.MarkdownNodeParser", "llama_index.indices.multi_modal.base.MultiModalVectorSto...
[((455, 524), 'llama_index.node_parser.MarkdownNodeParser', 'MarkdownNodeParser', ([], {'include_metadata': '(True)', 'include_prev_next_rel': '(True)'}), '(include_metadata=True, include_prev_next_rel=True)\n', (473, 524), False, 'from llama_index.node_parser import MarkdownNodeParser\n'), ((535, 579), 'openai.OpenAI'...
import streamlit as st import redirect as rd import os import tempfile import time from llama_index import StorageContext, LLMPredictor from llama_index import TreeIndex, load_index_from_storage from llama_index import ServiceContext from langchain.prompts import StringPromptTemplate from typing import List, Union fr...
[ "llama_index.ServiceContext.from_defaults", "llama_index.tools.ToolMetadata", "llama_index.StorageContext.from_defaults", "llama_index.load_index_from_storage", "llama_index.query_engine.MultiStepQueryEngine.from_defaults" ]
[((11873, 11906), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (11891, 11906), True, 'import streamlit as st\n'), ((11910, 11941), 'streamlit.title', 'st.title', (['"""Agriculture Web App"""'], {}), "('Agriculture Web App')\n", (11918, 11941), True, 'import str...
# Required Environment Variables: OPENAI_API_KEY # Required TavilyAI API KEY for web searches - https://tavily.com/ from llama_index.core import SimpleDirectoryReader from llama_index.packs.corrective_rag import CorrectiveRAGPack # load documents documents = SimpleDirectoryReader("./data").load_data() # uses the LLM ...
[ "llama_index.core.SimpleDirectoryReader", "llama_index.packs.corrective_rag.CorrectiveRAGPack" ]
[((387, 454), 'llama_index.packs.corrective_rag.CorrectiveRAGPack', 'CorrectiveRAGPack', (['documents'], {'tavily_ai_apikey': '"""<tavily_ai_apikey>"""'}), "(documents, tavily_ai_apikey='<tavily_ai_apikey>')\n", (404, 454), False, 'from llama_index.packs.corrective_rag import CorrectiveRAGPack\n'), ((260, 291), 'llama_...
# LLama Index starter example from: https://gpt-index.readthedocs.io/en/latest/getting_started/starter_example.html # In order to run this, download into data/ Paul Graham's Essay 'What I Worked On' from # https://github.com/jerryjliu/llama_index/blob/main/examples/paul_graham_essay/data/paul_graham_essay.txt # curl h...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.SimpleDirectoryReader", "llama_index.VectorStoreIndex", "llama_index.schema.TextNode", "llama_index.StorageContext.from_defaults", "llama_index.node_parser.SimpleNodeParser", "llama_index.schema.RelatedNodeInfo", "llama_index.load_index_from_...
[((789, 802), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (800, 802), False, 'from dotenv import load_dotenv\n'), ((809, 839), 'pprint.PrettyPrinter', 'pprint.PrettyPrinter', ([], {'indent': '(4)'}), '(indent=4)\n', (829, 839), False, 'import pprint\n'), ((970, 1012), 'llama_index.VectorStoreIndex.from_docum...
import argparse from pinecone import Pinecone from dotenv import load_dotenv import os from llama_index.vector_stores.pinecone import PineconeVectorStore from llama_index.core import VectorStoreIndex, StorageContext, ServiceContext from llama_index.embeddings.openai import OpenAIEmbedding from llama_index.readers.datab...
[ "llama_index.core.VectorStoreIndex.from_documents", "llama_index.core.StorageContext.from_defaults", "llama_index.vector_stores.pinecone.PineconeVectorStore", "llama_index.readers.database.DatabaseReader", "llama_index.core.ServiceContext.from_defaults", "llama_index.embeddings.openai.OpenAIEmbedding" ]
[((384, 397), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (395, 397), False, 'from dotenv import load_dotenv\n'), ((465, 520), 'llama_index.core.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {'vector_store': 'vector_store'}), '(vector_store=vector_store)\n', (493, 520), False, 'from lla...
import sounddevice as sd import wavio import whisper import openai from llama_index.llms import LlamaCPP from llama_index.llms.base import ChatMessage def record_audio(output_filename, duration, sample_rate): print("Recording...") audio_data = sd.rec(int(duration * sample_rate), ...
[ "llama_index.llms.LlamaCPP", "llama_index.llms.base.ChatMessage" ]
[((366, 375), 'sounddevice.wait', 'sd.wait', ([], {}), '()\n', (373, 375), True, 'import sounddevice as sd\n'), ((498, 564), 'wavio.write', 'wavio.write', (['output_filename', 'audio_data', 'sample_rate'], {'sampwidth': '(2)'}), '(output_filename, audio_data, sample_rate, sampwidth=2)\n', (509, 564), False, 'import wav...
import argparse import os from llama_index import StorageContext, load_index_from_storage from dotenv import load_dotenv from llama_index import VectorStoreIndex, SimpleDirectoryReader def query_data(query: str): """Query to a vector database ## argument Return: return_description """ sto...
[ "llama_index.load_index_from_storage", "llama_index.StorageContext.from_defaults" ]
[((335, 388), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {'persist_dir': '"""./storage"""'}), "(persist_dir='./storage')\n", (363, 388), False, 'from llama_index import StorageContext, load_index_from_storage\n'), ((418, 458), 'llama_index.load_index_from_storage', 'load_index_from...
from llama_index import SimpleDirectoryReader from llama_index import ServiceContext from langchain.chat_models import ChatOpenAI from llama_index import VectorStoreIndex from utils import build_sentence_window_index from utils import build_automerging_index import sys import os import logging import configparser c...
[ "llama_index.VectorStoreIndex.from_documents", "llama_index.ServiceContext.from_defaults", "llama_index.SimpleDirectoryReader" ]
[((328, 355), 'configparser.ConfigParser', 'configparser.ConfigParser', ([], {}), '()\n', (353, 355), False, 'import configparser\n'), ((2287, 2346), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.DEBUG'}), '(stream=sys.stdout, level=logging.DEBUG)\n', (2306, 2346), False,...
import streamlit as st import os import sys import openai from streamlit_extras.switch_page_button import switch_page from llama_index import VectorStoreIndex, GithubRepositoryReader, ServiceContext, set_global_service_context from llama_index.llms import OpenAI from database.neo4j_connection import connect_to_db st....
[ "llama_index.GithubRepositoryReader", "llama_index.ServiceContext.from_defaults", "llama_index.llms.OpenAI", "llama_index.set_global_service_context" ]
[((317, 379), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Authentication"""', 'page_icon': '"""🔐"""'}), "(page_title='Authentication', page_icon='🔐')\n", (335, 379), True, 'import streamlit as st\n'), ((392, 439), 'streamlit.write', 'st.write', (['"""# Welcome to AI GitHub Repo reader!"...
from llama_index.core import VectorStoreIndex,SimpleDirectoryReader,ServiceContext print("VectorStoreIndex,SimpleDirectoryReader,ServiceContext imported") from llama_index.llms.huggingface import HuggingFaceLLM print("HuggingFaceLLM imported") from llama_index.core.prompts.prompts import SimpleInputPrompt print("Simple...
[ "llama_index.core.SimpleDirectoryReader", "llama_index.core.prompts.prompts.SimpleInputPrompt" ]
[((872, 885), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (883, 885), False, 'from dotenv import load_dotenv\n'), ((905, 931), 'os.environ.get', 'os.environ.get', (['"""HF_TOKEN"""'], {}), "('HF_TOKEN')\n", (919, 931), False, 'import os\n'), ((1262, 1315), 'llama_index.core.prompts.prompts.SimpleInputPrompt'...
import logging from typing import Any, List, Optional from llama_index.legacy.bridge.pydantic import Field, PrivateAttr from llama_index.legacy.core.embeddings.base import ( DEFAULT_EMBED_BATCH_SIZE, BaseEmbedding, Embedding, ) logger = logging.getLogger(__name__) # For bge models that Gradient AI provi...
[ "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.bridge.pydantic.Field" ]
[((251, 278), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (268, 278), False, 'import logging\n'), ((1040, 1086), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'default': 'GRADIENT_EMBED_BATCH_SIZE', 'gt': '(0)'}), '(default=GRADIENT_EMBED_BATCH_SIZE, gt=0)\n', (1045, 1086)...
"""Answer inserter.""" from abc import abstractmethod from typing import Any, Dict, List, Optional from llama_index.core.llms.llm import LLM from llama_index.core.prompts.base import BasePromptTemplate, PromptTemplate from llama_index.core.prompts.mixin import ( PromptDictType, PromptMixin, PromptMixinTyp...
[ "llama_index.core.prompts.base.PromptTemplate", "llama_index.core.settings.llm_from_settings_or_context" ]
[((4287, 4336), 'llama_index.core.prompts.base.PromptTemplate', 'PromptTemplate', (['DEFAULT_ANSWER_INSERT_PROMPT_TMPL'], {}), '(DEFAULT_ANSWER_INSERT_PROMPT_TMPL)\n', (4301, 4336), False, 'from llama_index.core.prompts.base import BasePromptTemplate, PromptTemplate\n'), ((4860, 4915), 'llama_index.core.settings.llm_fr...
"""Retrieval evaluators.""" from typing import Any, List, Optional, Sequence, Tuple from llama_index.legacy.bridge.pydantic import Field from llama_index.legacy.core.base_retriever import BaseRetriever from llama_index.legacy.evaluation.retrieval.base import ( BaseRetrievalEvaluator, RetrievalEvalMode, ) from...
[ "llama_index.legacy.bridge.pydantic.Field" ]
[((1038, 1085), 'llama_index.legacy.bridge.pydantic.Field', 'Field', (['...'], {'description': '"""Retriever to evaluate"""'}), "(..., description='Retriever to evaluate')\n", (1043, 1085), False, 'from llama_index.legacy.bridge.pydantic import Field\n'), ((1151, 1209), 'llama_index.legacy.bridge.pydantic.Field', 'Fiel...
from typing import Any, List, Optional from llama_index.legacy.bridge.pydantic import Field, PrivateAttr from llama_index.legacy.callbacks import CallbackManager from llama_index.legacy.core.embeddings.base import ( DEFAULT_EMBED_BATCH_SIZE, BaseEmbedding, ) from llama_index.legacy.embeddings.huggingface_utils...
[ "llama_index.legacy.embeddings.huggingface_utils.format_query", "llama_index.legacy.embeddings.huggingface_utils.get_pooling_mode", "llama_index.legacy.embeddings.pooling.Pooling", "llama_index.legacy.bridge.pydantic.PrivateAttr", "llama_index.legacy.bridge.pydantic.Field", "llama_index.legacy.embeddings....
[((567, 613), 'llama_index.legacy.bridge.pydantic.Field', 'Field', ([], {'description': '"""Folder name to load from."""'}), "(description='Folder name to load from.')\n", (572, 613), False, 'from llama_index.legacy.bridge.pydantic import Field, PrivateAttr\n'), ((636, 681), 'llama_index.legacy.bridge.pydantic.Field', ...
"""LLM Chains for executing Retrival Augmented Generation.""" import base64 import os from functools import lru_cache from pathlib import Path from typing import TYPE_CHECKING, Generator, List, Optional import torch from langchain.embeddings import HuggingFaceEmbeddings from langchain.llms import HuggingFaceTextGenInf...
[ "llama_index.download_loader", "llama_index.vector_stores.MilvusVectorStore", "llama_index.embeddings.LangchainEmbedding", "llama_index.llms.LangChainLLM", "llama_index.Prompt", "llama_index.node_parser.SimpleNodeParser.from_defaults", "llama_index.set_global_service_context", "llama_index.response.sc...
[((3156, 3202), 'os.environ.get', 'os.environ.get', (['"""APP_CONFIG_FILE"""', '"""/dev/null"""'], {}), "('APP_CONFIG_FILE', '/dev/null')\n", (3170, 3202), False, 'import os\n'), ((3216, 3262), 'chain_server.configuration.AppConfig.from_file', 'configuration.AppConfig.from_file', (['config_file'], {}), '(config_file)\n...
from llama_index import ServiceContext from llama_index import StorageContext, load_index_from_storage from omegaconf import DictConfig, OmegaConf import hydra from llama_index.evaluation import RetrieverEvaluator from llama_index.evaluation import ( EmbeddingQAFinetuneDataset, ) import pandas as pd @hydra.main(v...
[ "llama_index.evaluation.RetrieverEvaluator.from_metric_names", "llama_index.ServiceContext.from_defaults", "llama_index.StorageContext.from_defaults", "llama_index.evaluation.EmbeddingQAFinetuneDataset.from_json", "llama_index.load_index_from_storage" ]
[((308, 385), 'hydra.main', 'hydra.main', ([], {'version_base': 'None', 'config_path': '"""../../conf"""', 'config_name': '"""config"""'}), "(version_base=None, config_path='../../conf', config_name='config')\n", (318, 385), False, 'import hydra\n'), ((589, 619), 'llama_index.ServiceContext.from_defaults', 'ServiceCont...
import os import time from typing import Any, Callable, List, Sequence from lib import constants from lib.index.helper import cur_simple_date_time_sec from llama_index.core.llms.callbacks import llm_chat_callback, llm_completion_callback from llama_index.core.base.llms.base import BaseLLM from llama_index.core.llms imp...
[ "llama_index.core.llms.callbacks.llm_completion_callback", "llama_index.core.llms.callbacks.llm_chat_callback" ]
[((1905, 1924), 'llama_index.core.llms.callbacks.llm_chat_callback', 'llm_chat_callback', ([], {}), '()\n', (1922, 1924), False, 'from llama_index.core.llms.callbacks import llm_chat_callback, llm_completion_callback\n'), ((2715, 2734), 'llama_index.core.llms.callbacks.llm_chat_callback', 'llm_chat_callback', ([], {}),...
from llama_index import StorageContext, load_index_from_storage # rebuild storage context storage_context = StorageContext.from_defaults(persist_dir="./storage") # load index index = load_index_from_storage(storage_context)
[ "llama_index.load_index_from_storage", "llama_index.StorageContext.from_defaults" ]
[((109, 162), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {'persist_dir': '"""./storage"""'}), "(persist_dir='./storage')\n", (137, 162), False, 'from llama_index import StorageContext, load_index_from_storage\n'), ((184, 224), 'llama_index.load_index_from_storage', 'load_index_from...
import os from dotenv import load_dotenv from llama_index import PromptTemplate, SimpleDirectoryReader, VectorStoreIndex from ragas.metrics import ( faithfulness, answer_relevancy, context_precision, context_recall, ) from ragas.metrics.critique import harmfulness from ragas.llama_index import evaluate...
[ "llama_index.SimpleDirectoryReader", "llama_index.evaluation.QueryResponseDataset.from_json", "llama_index.PromptTemplate", "llama_index.VectorStoreIndex.from_vector_store", "llama_index.node_parser.SentenceSplitter" ]
[((1361, 1411), 'llama_index.node_parser.SentenceSplitter', 'SentenceSplitter', ([], {'chunk_size': '(256)', 'chunk_overlap': '(50)'}), '(chunk_size=256, chunk_overlap=50)\n', (1377, 1411), False, 'from llama_index.node_parser import SentenceSplitter\n'), ((1440, 1491), 'llama_index.node_parser.SentenceSplitter', 'Sent...
from __future__ import annotations from typing import TYPE_CHECKING, List import logging import json import commentjson as cjson import os import sys import requests import urllib3 from tqdm import tqdm import colorama from duckduckgo_search import ddg import asyncio import aiohttp from enum import Enum from .preset...
[ "llama_index.ServiceContext.from_defaults", "llama_index.OpenAIEmbedding", "llama_index.indices.vector_store.base_query.GPTVectorStoreIndexQuery", "llama_index.indices.query.schema.QueryBundle" ]
[((2508, 2593), 'logging.warning', 'logging.warning', (['"""stream predict not implemented, using at once predict instead"""'], {}), "('stream predict not implemented, using at once predict instead'\n )\n", (2523, 2593), False, 'import logging\n'), ((2944, 3029), 'logging.warning', 'logging.warning', (['"""at once p...
"""Autoretriever prompts.""" from llama_index.legacy.prompts.base import PromptTemplate from llama_index.legacy.prompts.prompt_type import PromptType from llama_index.legacy.vector_stores.types import ( FilterOperator, MetadataFilter, MetadataInfo, VectorStoreInfo, VectorStoreQuerySpec, ) # NOTE: ...
[ "llama_index.legacy.vector_stores.types.MetadataInfo", "llama_index.legacy.vector_stores.types.MetadataFilter", "llama_index.legacy.prompts.base.PromptTemplate" ]
[((3927, 4033), 'llama_index.legacy.prompts.base.PromptTemplate', 'PromptTemplate', ([], {'template': 'DEFAULT_VECTARA_QUERY_PROMPT_TMPL', 'prompt_type': 'PromptType.VECTOR_STORE_QUERY'}), '(template=DEFAULT_VECTARA_QUERY_PROMPT_TMPL, prompt_type=\n PromptType.VECTOR_STORE_QUERY)\n', (3941, 4033), False, 'from llama...
import asyncio from abc import abstractmethod from typing import Any, Dict, List, Optional, Sequence, Tuple, cast import pandas as pd from tqdm import tqdm from llama_index.core.async_utils import DEFAULT_NUM_WORKERS, run_jobs from llama_index.core.base.response.schema import PydanticResponse from llama_index.core.br...
[ "llama_index.core.service_context.ServiceContext.from_defaults", "llama_index.llms.openai.OpenAI", "llama_index.core.bridge.pydantic.Field", "llama_index.core.node_parser.SentenceSplitter", "llama_index.core.callbacks.base.CallbackManager", "llama_index.core.async_utils.run_jobs", "llama_index.core.sche...
[((2154, 2206), 'llama_index.core.bridge.pydantic.Field', 'Field', ([], {'default_factory': 'CallbackManager', 'exclude': '(True)'}), '(default_factory=CallbackManager, exclude=True)\n', (2159, 2206), False, 'from llama_index.core.bridge.pydantic import BaseModel, Field, ValidationError\n'), ((2246, 2316), 'llama_index...
"""Faithfulness evaluation.""" from __future__ import annotations from typing import Any, List, Optional, Sequence, Union from llama_index.core.evaluation.base import BaseEvaluator, EvaluationResult from llama_index.core.multi_modal_llms.base import MultiModalLLM from llama_index.core.prompts import BasePromptTempla...
[ "llama_index.core.prompts.PromptTemplate", "llama_index.core.evaluation.base.EvaluationResult", "llama_index.core.schema.ImageNode", "llama_index.multi_modal_llms.openai.OpenAIMultiModal" ]
[((468, 1757), 'llama_index.core.prompts.PromptTemplate', 'PromptTemplate', (['"""Please tell if a given piece of information is supported by the visual as well as textual context information.\nYou need to answer with either YES or NO.\nAnswer YES if any of the image(s) and textual context supports the information, eve...
# SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. # SPDX-License-Identifier: MIT # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the "Software"), # to deal in the Software without res...
[ "llama_index.llms.base.llm_chat_callback", "llama_index.llms.base.LLMMetadata", "llama_index.bridge.pydantic.Field", "llama_index.llms.generic_utils.completion_response_to_chat_response", "llama_index.bridge.pydantic.PrivateAttr", "llama_index.llms.base.llm_completion_callback" ]
[((2151, 2199), 'llama_index.bridge.pydantic.Field', 'Field', ([], {'description': '"""The path to the trt engine."""'}), "(description='The path to the trt engine.')\n", (2156, 2199), False, 'from llama_index.bridge.pydantic import Field, PrivateAttr\n'), ((2239, 2296), 'llama_index.bridge.pydantic.Field', 'Field', ([...
from llama_index.core.callbacks.schema import CBEventType, EventPayload from llama_index.core.llms import ChatMessage, ChatResponse from llama_index.core.schema import NodeWithScore, TextNode import chainlit as cl @cl.on_chat_start async def start(): await cl.Message(content="LlamaIndexCb").send() cb = cl.L...
[ "llama_index.core.schema.TextNode", "llama_index.core.llms.ChatMessage" ]
[((316, 346), 'chainlit.LlamaIndexCallbackHandler', 'cl.LlamaIndexCallbackHandler', ([], {}), '()\n', (344, 346), True, 'import chainlit as cl\n'), ((415, 428), 'chainlit.sleep', 'cl.sleep', (['(0.2)'], {}), '(0.2)\n', (423, 428), True, 'import chainlit as cl\n'), ((691, 704), 'chainlit.sleep', 'cl.sleep', (['(0.2)'], ...