code
stringlengths
141
78.9k
apis
listlengths
1
23
extract_api
stringlengths
142
73.2k
import os import time import openai import pickle import langchain import streamlit as st from langchain import OpenAI from langchain.chains import RetrievalQAWithSourcesChain from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.document_loaders import UnstructuredURLLoader from langchain.e...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.UnstructuredURLLoader", "langchain.vectorstores.FAISS.from_documents", "langchain.embeddings.OpenAIEmbeddings", "langchain.OpenAI" ]
[((487, 518), 'streamlit.title', 'st.title', (['"""News Research tool """'], {}), "('News Research tool ')\n", (495, 518), True, 'import streamlit as st\n'), ((519, 556), 'streamlit.sidebar.title', 'st.sidebar.title', (['"""News Article URLs"""'], {}), "('News Article URLs')\n", (535, 556), True, 'import streamlit as s...
""" This example shows how to use the map-reduce chain to summarize a document. """ import os import langchain from langchain_openai import ChatOpenAI from langchain.chains.summarize import load_summarize_chain from langchain_community.document_loaders import PyPDFLoader from dotenv import load_dotenv lo...
[ "langchain_community.document_loaders.PyPDFLoader", "langchain_openai.ChatOpenAI", "langchain.chains.summarize.load_summarize_chain" ]
[((318, 331), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (329, 331), False, 'from dotenv import load_dotenv\n'), ((352, 379), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (361, 379), False, 'import os\n'), ((415, 479), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'...
"""LLM Chains for executing Retrival Augmented Generation.""" import base64 import os from functools import lru_cache from pathlib import Path from typing import TYPE_CHECKING, Generator, List, Optional import torch from langchain.embeddings import HuggingFaceEmbeddings from langchain.llms import HuggingFaceTextGenInf...
[ "langchain.llms.HuggingFaceTextGenInference", "langchain.embeddings.HuggingFaceEmbeddings" ]
[((3156, 3202), 'os.environ.get', 'os.environ.get', (['"""APP_CONFIG_FILE"""', '"""/dev/null"""'], {}), "('APP_CONFIG_FILE', '/dev/null')\n", (3170, 3202), False, 'import os\n'), ((3216, 3262), 'chain_server.configuration.AppConfig.from_file', 'configuration.AppConfig.from_file', (['config_file'], {}), '(config_file)\n...
from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext from llama_index import LangchainEmbedding from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_setup import llm def setup_memory(): documents = SimpleDirectoryReader("./Data").load_data() embed_model = Lan...
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((429, 507), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'chunk_size': '(256)', 'llm': 'llm', 'embed_model': 'embed_model'}), '(chunk_size=256, llm=llm, embed_model=embed_model)\n', (457, 507), False, 'from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext...
from modules.preprocessors import BasePreprocessor from modules.templates import CONDENSE_QUESTION_TEMPLATE from utils import create_collection, create_save_collection import langchain from typing import Optional, Any, Dict, Union from langchain.schema import BaseDocumentTransformer from langchain.schema.prompt_templa...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI", "langchain.cache.InMemoryCache", "langchain.vectorstores.Chroma" ]
[((1674, 1689), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (1687, 1689), False, 'from langchain.cache import InMemoryCache\n'), ((3798, 3896), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'output_key': '"""answer"""', 'return...
import os from dotenv import load_dotenv, find_dotenv _ = load_dotenv(find_dotenv()) # read local .env file from langchain.chains import RetrievalQA from langchain.chat_models import ChatOpenAI from langchain.document_loaders import CSVLoader from langchain.indexes import VectorstoreIndexCreator from langchain.vecto...
[ "langchain.evaluation.qa.QAEvalChain.from_llm", "langchain.document_loaders.CSVLoader", "langchain.indexes.VectorstoreIndexCreator", "langchain.chat_models.ChatOpenAI" ]
[((409, 434), 'langchain.document_loaders.CSVLoader', 'CSVLoader', ([], {'file_path': 'file'}), '(file_path=file)\n', (418, 434), False, 'from langchain.document_loaders import CSVLoader\n'), ((565, 592), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0.0)'}), '(temperature=0.0)\n', (575, 592)...
import langchain_visualizer # isort:skip # noqa: F401 from fvalues import FValue from langchain import FewShotPromptTemplate, PromptTemplate def test_few_shot_f(): examples = [ {"word": "happy", "antonym": "sad"}, {"word": "tall", "antonym": "short"}, # Should be able to handle extra ke...
[ "langchain.FewShotPromptTemplate", "langchain.PromptTemplate" ]
[((455, 544), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['word', 'antonym']", 'template': '"""w={word},a={antonym}"""'}), "(input_variables=['word', 'antonym'], template=\n 'w={word},a={antonym}')\n", (469, 544), False, 'from langchain import FewShotPromptTemplate, PromptTemplate\n'), (...
import langchain.utilities.opaqueprompts as op from langchain import LLMChain, PromptTemplate from langchain.llms import OpenAI from langchain.llms.opaqueprompts import OpaquePrompts from langchain.memory import ConversationBufferWindowMemory from langchain.schema.output_parser import StrOutputParser from langchain.sch...
[ "langchain.memory.ConversationBufferWindowMemory", "langchain.schema.output_parser.StrOutputParser", "langchain.llms.OpenAI", "langchain.PromptTemplate.from_template", "langchain.utilities.opaqueprompts.desanitize" ]
[((2863, 2871), 'langchain.llms.OpenAI', 'OpenAI', ([], {}), '()\n', (2869, 2871), False, 'from langchain.llms import OpenAI\n'), ((2805, 2850), 'langchain.PromptTemplate.from_template', 'PromptTemplate.from_template', (['prompt_template'], {}), '(prompt_template)\n', (2833, 2850), False, 'from langchain import LLMChai...
from langchain.chat_models import ChatOpenAI from dreamsboard.dreams.dreams_personality_chain.base import StoryBoardDreamsGenerationChain import logging import langchain langchain.verbose = True logger = logging.getLogger(__name__) logger.setLevel(logging.INFO) # 控制台打印 handler = logging.StreamHandler() handler.setLev...
[ "langchain.chat_models.ChatOpenAI" ]
[((205, 232), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (222, 232), False, 'import logging\n'), ((282, 305), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (303, 305), False, 'import logging\n'), ((782, 806), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([],...
"""Test Upstash Redis cache functionality.""" import uuid import pytest import langchain from langchain.cache import UpstashRedisCache from langchain.schema import Generation, LLMResult from tests.unit_tests.llms.fake_chat_model import FakeChatModel from tests.unit_tests.llms.fake_llm import FakeLLM URL = "<UPSTASH_...
[ "langchain.llm_cache.clear", "langchain.schema.Generation", "langchain.llm_cache.redis.flushall", "langchain.llm_cache.redis.pttl", "langchain.llm_cache._key", "langchain.llm_cache.lookup" ]
[((436, 473), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (456, 473), False, 'import pytest\n'), ((809, 846), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (829, 846), False, 'import pytest\n'), ((2491, 252...
from uuid import UUID from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser, initialize_agent from langchain.prompts import HumanMessagePromptTemplate, SystemMessagePromptTemplate, ChatPromptTemplate, AIMessagePromptTemplate, PromptTemplate from langchain import OpenAI, SerpAPIWrappe...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.agents.LLMSingleActionAgent", "langchain.LLMChain", "langchain.schema.AgentAction", "langchain.chat_models.ChatOpenAI", "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.prompts.HumanMessagePromptTemplate", "langchain.sch...
[((987, 1047), 'langchain.prompts.SystemMessagePromptTemplate', 'SystemMessagePromptTemplate', ([], {'prompt': 'test_human_system_prompt'}), '(prompt=test_human_system_prompt)\n', (1014, 1047), False, 'from langchain.prompts import HumanMessagePromptTemplate, SystemMessagePromptTemplate, ChatPromptTemplate, AIMessagePr...
import time import unittest.mock from typing import Any from uuid import UUID from langchainplus_sdk import LangChainPlusClient from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.callbacks.tracers.schemas import Run from langchain.schema.output import LLMResult def test_example_id_assi...
[ "langchain.callbacks.tracers.langchain.LangChainTracer", "langchainplus_sdk.LangChainPlusClient", "langchain.schema.output.LLMResult" ]
[((741, 762), 'langchainplus_sdk.LangChainPlusClient', 'LangChainPlusClient', ([], {}), '()\n', (760, 762), False, 'from langchainplus_sdk import LangChainPlusClient\n'), ((780, 810), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (795, 810),...
# Copyright (c) Meta Platforms, Inc. and affiliates. # This software may be used and distributed according to the terms of the Llama 2 Community License Agreement. import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json os.environ[...
[ "langchain.llms.Replicate" ]
[((488, 595), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llama2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llama2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (497, 595), False, 'from langchain.llms import R...
"""**Document Transformers** are classes to transform Documents. **Document Transformers** usually used to transform a lot of Documents in a single run. **Class hierarchy:** .. code-block:: BaseDocumentTransformer --> <name> # Examples: DoctranQATransformer, DoctranTextTranslator **Main helpers:** .. code-bl...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((677, 697), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (695, 697), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((707, 1102), 'warnings.warn', 'warnings.warn', (['f"""Importing document transformers from langchain is deprecated. Importi...
"""**Document Transformers** are classes to transform Documents. **Document Transformers** usually used to transform a lot of Documents in a single run. **Class hierarchy:** .. code-block:: BaseDocumentTransformer --> <name> # Examples: DoctranQATransformer, DoctranTextTranslator **Main helpers:** .. code-bl...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((677, 697), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (695, 697), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((707, 1102), 'warnings.warn', 'warnings.warn', (['f"""Importing document transformers from langchain is deprecated. Importi...
"""**Document Transformers** are classes to transform Documents. **Document Transformers** usually used to transform a lot of Documents in a single run. **Class hierarchy:** .. code-block:: BaseDocumentTransformer --> <name> # Examples: DoctranQATransformer, DoctranTextTranslator **Main helpers:** .. code-bl...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((677, 697), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (695, 697), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((707, 1102), 'warnings.warn', 'warnings.warn', (['f"""Importing document transformers from langchain is deprecated. Importi...
"""**Document Transformers** are classes to transform Documents. **Document Transformers** usually used to transform a lot of Documents in a single run. **Class hierarchy:** .. code-block:: BaseDocumentTransformer --> <name> # Examples: DoctranQATransformer, DoctranTextTranslator **Main helpers:** .. code-bl...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((677, 697), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (695, 697), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((707, 1102), 'warnings.warn', 'warnings.warn', (['f"""Importing document transformers from langchain is deprecated. Importi...
"""Beta Feature: base interface for cache.""" from __future__ import annotations import hashlib import inspect import json import logging from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, ...
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((918, 945), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (935, 945), False, 'import logging\n'), ((3390, 3408), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3406, 3408), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((356...
"""Beta Feature: base interface for cache.""" from __future__ import annotations import hashlib import inspect import json import logging from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, ...
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((918, 945), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (935, 945), False, 'import logging\n'), ((3390, 3408), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3406, 3408), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((356...
"""Beta Feature: base interface for cache.""" from __future__ import annotations import hashlib import inspect import json import logging from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, ...
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((918, 945), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (935, 945), False, 'import logging\n'), ((3390, 3408), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3406, 3408), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((356...
# Import Langchain modules from langchain.document_loaders import PyPDFLoader from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import FAISS from langchain.chains import RetrievalQA from langchain.llms import OpenAI # Impo...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.vectorstores.FAISS.from_documents", "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.OpenAIEmbeddings" ]
[((573, 606), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (596, 606), False, 'import warnings\n'), ((712, 808), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)s - %(levelname)s - %(message)s"""'}), "(level=logging.IN...
import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class WhatsAppClient: API_URL = "https://graph.facebook.com/v17.0/" WHATSAPP_API_TOKEN = "<Temporary access token from your WhatsApp API Setup>" WHATSAPP_CLOUD_NUM...
[ "langchain.llms.Replicate" ]
[((1337, 1444), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llama2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llama2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1346, 1444), False, 'from langchain.llms impo...
import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class WhatsAppClient: API_URL = "https://graph.facebook.com/v17.0/" WHATSAPP_API_TOKEN = "<Temporary access token from your WhatsApp API Setup>" WHATSAPP_CLOUD_NUM...
[ "langchain.llms.Replicate" ]
[((1337, 1444), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llama2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llama2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1346, 1444), False, 'from langchain.llms impo...
import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class WhatsAppClient: API_URL = "https://graph.facebook.com/v17.0/" WHATSAPP_API_TOKEN = "<Temporary access token from your WhatsApp API Setup>" WHATSAPP_CLOUD_NUM...
[ "langchain.llms.Replicate" ]
[((1337, 1444), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llama2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llama2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1346, 1444), False, 'from langchain.llms impo...
import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class WhatsAppClient: API_URL = "https://graph.facebook.com/v17.0/" WHATSAPP_API_TOKEN = "<Temporary access token from your WhatsApp API Setup>" WHATSAPP_CLOUD_NUM...
[ "langchain.llms.Replicate" ]
[((1337, 1444), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llama2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llama2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1346, 1444), False, 'from langchain.llms impo...
"""Utility functions for mlflow.langchain.""" import json import logging import os import shutil import types from functools import lru_cache from importlib.util import find_spec from typing import NamedTuple import cloudpickle import yaml from packaging import version import mlflow from mlflow.utils.class_utils impo...
[ "langchain.schema.output_parser.StrOutputParser", "langchain.agents.initialize_agent", "langchain.chains.loading.load_chain" ]
[((2001, 2028), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (2018, 2028), False, 'import logging\n'), ((10189, 10235), 'os.path.join', 'os.path.join', (['path', '_MODEL_DATA_YAML_FILE_NAME'], {}), '(path, _MODEL_DATA_YAML_FILE_NAME)\n', (10201, 10235), False, 'import os\n'), ((5685, 57...
"""Beta Feature: base interface for cache.""" import hashlib import json from abc import ABC, abstractmethod from typing import Any, Callable, Dict, List, Optional, Tuple, Type, cast from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm impo...
[ "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.schema.Generation" ]
[((2037, 2055), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (2053, 2055), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((2212, 2244), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (2218, 2244), Fal...
# Needs to be in same directory as configs, data folder # Imports from _OpalLLM import OpalLLM from _OpalLLM import OpalLLM import sys sys.path.append('/home/jovyan/.local/lib/python3.8/site-packages') import torch from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langcha...
[ "langchain.chains.ConversationChain", "langchain.LLMChain", "langchain.llms.HuggingFacePipeline", "langchain.PromptTemplate" ]
[((138, 204), 'sys.path.append', 'sys.path.append', (['"""/home/jovyan/.local/lib/python3.8/site-packages"""'], {}), "('/home/jovyan/.local/lib/python3.8/site-packages')\n", (153, 204), False, 'import sys\n'), ((3396, 3513), '_OpalLLM.OpalLLM', 'OpalLLM', ([], {'model': '"""lmsys/vicuna-33b"""', 'temperature': '(0.1)',...
"""Beta Feature: base interface for cache.""" from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequenc...
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((950, 977), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (967, 977), False, 'import logging\n'), ((3422, 3440), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3438, 3440), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((359...
"""Beta Feature: base interface for cache.""" from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequenc...
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((950, 977), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (967, 977), False, 'import logging\n'), ((3422, 3440), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3438, 3440), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((359...
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.bas...
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages....
[((915, 952), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (920, 952), False, 'from pydantic import Field, root_validator\n'), ((1026, 1059), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (103...
from __future__ import annotations import asyncio import functools import logging import os import uuid from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( TYPE_CHECKING, Any, AsyncGenerator, Dict, Generator, List, Optional, ...
[ "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.l...
[((1329, 1356), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1346, 1356), False, 'import logging\n'), ((1425, 1468), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1435, 1468), False, 'from contextvars i...
from __future__ import annotations import asyncio import functools import logging import os import uuid from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( TYPE_CHECKING, Any, AsyncGenerator, Dict, Generator, List, Optional, ...
[ "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.l...
[((1329, 1356), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1346, 1356), False, 'import logging\n'), ((1425, 1468), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1435, 1468), False, 'from contextvars i...
from __future__ import annotations import asyncio import functools import logging import os import uuid from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( TYPE_CHECKING, Any, AsyncGenerator, Dict, Generator, List, Optional, ...
[ "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.l...
[((1329, 1356), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1346, 1356), False, 'import logging\n'), ((1425, 1468), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1435, 1468), False, 'from contextvars i...
from __future__ import annotations import asyncio import functools import logging import os import uuid from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( TYPE_CHECKING, Any, AsyncGenerator, Dict, Generator, List, Optional, ...
[ "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.l...
[((1329, 1356), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1346, 1356), False, 'import logging\n'), ((1425, 1468), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1435, 1468), False, 'from contextvars i...
from modules.preprocessors import BasePreprocessor from modules.templates import CONDENSE_QUESTION_TEMPLATE from utils import create_collection, create_save_collection import langchain from typing import Optional, Any, Dict, Union from langchain.schema import BaseDocumentTransformer from langchain.schema.prompt_templa...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI", "langchain.cache.InMemoryCache", "langchain.vectorstores.Chroma" ]
[((1674, 1689), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (1687, 1689), False, 'from langchain.cache import InMemoryCache\n'), ((3798, 3896), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'output_key': '"""answer"""', 'return...
from modules.preprocessors import BasePreprocessor from modules.templates import CONDENSE_QUESTION_TEMPLATE from utils import create_collection, create_save_collection import langchain from typing import Optional, Any, Dict, Union from langchain.schema import BaseDocumentTransformer from langchain.schema.prompt_templa...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI", "langchain.cache.InMemoryCache", "langchain.vectorstores.Chroma" ]
[((1674, 1689), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (1687, 1689), False, 'from langchain.cache import InMemoryCache\n'), ((3798, 3896), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'output_key': '"""answer"""', 'return...
"""Test Upstash Redis cache functionality.""" import uuid import pytest import langchain from langchain.cache import UpstashRedisCache from langchain.schema import Generation, LLMResult from tests.unit_tests.llms.fake_chat_model import FakeChatModel from tests.unit_tests.llms.fake_llm import FakeLLM URL = "<UPSTASH_...
[ "langchain.llm_cache.clear", "langchain.schema.Generation", "langchain.llm_cache.redis.flushall", "langchain.llm_cache.redis.pttl", "langchain.llm_cache._key", "langchain.llm_cache.lookup" ]
[((436, 473), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (456, 473), False, 'import pytest\n'), ((809, 846), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (829, 846), False, 'import pytest\n'), ((2491, 252...
"""Test Upstash Redis cache functionality.""" import uuid import pytest import langchain from langchain.cache import UpstashRedisCache from langchain.schema import Generation, LLMResult from tests.unit_tests.llms.fake_chat_model import FakeChatModel from tests.unit_tests.llms.fake_llm import FakeLLM URL = "<UPSTASH_...
[ "langchain.llm_cache.clear", "langchain.schema.Generation", "langchain.llm_cache.redis.flushall", "langchain.llm_cache.redis.pttl", "langchain.llm_cache._key", "langchain.llm_cache.lookup" ]
[((436, 473), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (456, 473), False, 'import pytest\n'), ((809, 846), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (829, 846), False, 'import pytest\n'), ((2491, 252...
''' Create Vector Store from all documents in a folder, currently supports .pptx, .docx, .pdf files. Created by Ric Zhou on 2021-03-27 ''' from langchain.document_loaders import (UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredPDFLoader) import glob import langchain.text_splitte...
[ "langchain.document_loaders.UnstructuredWordDocumentLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.UnstructuredPowerPointLoader", "langchain.vectorstores.FAISS.save_local", "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.OpenAIEmbeddings" ]
[((604, 617), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (615, 617), False, 'from dotenv import load_dotenv\n'), ((618, 633), 'GlobalClasses.GlobalContext', 'GlobalContext', ([], {}), '()\n', (631, 633), False, 'from GlobalClasses import GlobalContext\n'), ((939, 1017), 'langchain.text_splitter.RecursiveCha...
import os import key import tabulate # Set API key os.environ["OPENAI_API_KEY"] = key.OPENAI_API_KEY # Import langchain from langchain.chains import RetrievalQA from langchain.chat_models import ChatOpenAI from langchain.document_loaders import CSVLoader from langchain.indexes import VectorstoreIndexCreator from langc...
[ "langchain.document_loaders.CSVLoader", "langchain.indexes.VectorstoreIndexCreator", "langchain.chat_models.ChatOpenAI" ]
[((465, 508), 'langchain.document_loaders.CSVLoader', 'CSVLoader', ([], {'file_path': 'file', 'encoding': '"""utf-8"""'}), "(file_path=file, encoding='utf-8')\n", (474, 508), False, 'from langchain.document_loaders import CSVLoader\n'), ((708, 735), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature':...
import openai import langchain as lc from langchain.llms import OpenAI import gradio as gr # 设置OpenAI API密钥 openai.api_key = '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' # 初始化Langchain的OpenAI LLM llm = OpenAI(api_key=openai.api_key) # 定义一个函数来处理上传的文档并生成响应 def process_document(document): # 这里可以添加代码来处理文档,...
[ "langchain.llms.OpenAI" ]
[((213, 243), 'langchain.llms.OpenAI', 'OpenAI', ([], {'api_key': 'openai.api_key'}), '(api_key=openai.api_key)\n', (219, 243), False, 'from langchain.llms import OpenAI\n'), ((508, 536), 'gradio.inputs.File', 'gr.inputs.File', ([], {'label': '"""上传文档"""'}), "(label='上传文档')\n", (522, 536), True, 'import gradio as gr\n'...
import os import pandas as pd import math from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain.text_splitter import CharacterTextSplitter from langchain import OpenAI, VectorDBQA, OpenAI from langchain.llms import OpenAIChat from langchain.document_loaders im...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.vectorstores.Chroma.from_documents", "langchain.document_loaders.DataFrameLoader", "langchain.OpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((527, 555), 'sys.modules.pop', 'sys.modules.pop', (['"""pysqlite3"""'], {}), "('pysqlite3')\n", (542, 555), False, 'import sys\n'), ((558, 587), 'streamlit.title', 'st.title', (['"""GPT module (TEST)"""'], {}), "('GPT module (TEST)')\n", (566, 587), True, 'import streamlit as st\n'), ((606, 660), 'streamlit.text_inpu...
# Python built-in module import os import time import json # Python installed module import tiktoken import langchain from spacy.lang.en import English class SentencizerSplitter(object): def __init__(self, config_dict): self.total_tokens = config_dict["embedding"]["total_tokens"] self.approx_tota...
[ "langchain.schema.document.Document" ]
[((511, 520), 'spacy.lang.en.English', 'English', ([], {}), '()\n', (518, 520), False, 'from spacy.lang.en import English\n'), ((586, 653), 'tiktoken.encoding_for_model', 'tiktoken.encoding_for_model', (["config_dict['embedding']['model_name']"], {}), "(config_dict['embedding']['model_name'])\n", (613, 653), False, 'im...
import os import json import openai from utils import * import random import langchain from langchain import PromptTemplate from langchain.llms import OpenAI, OpenAIChat from langchain import LLMChain from re import compile from datetime import datetime from typing import NamedTuple from openai import Embedding #set ...
[ "langchain.llms.OpenAIChat", "langchain.LLMChain", "langchain.PromptTemplate" ]
[((1869, 1883), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1881, 1883), False, 'from datetime import datetime\n'), ((2826, 2890), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_text', 'input_variables': "['Memory']"}), "(template=prompt_text, input_variables=['Memory'])\n", (28...
# Copyright (c) Khulnasoft Platforms, Inc. and affiliates. # This software may be used and distributed according to the terms of the Llmk 2 Community License Agreement. import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class ...
[ "langchain.llms.Replicate" ]
[((1513, 1619), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llmk2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llmk2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1522, 1619), False, 'from langchain.llms import...
import os import langchain from config import * from util import * from langchain.llms import OpenAI, Cohere, HuggingFaceHub from langchain.chat_models import ChatOpenAI from langchain.agents import AgentType, initialize_agent, load_tools from typing import Optional, Type from langchain.callbacks.manager import AsyncCa...
[ "langchain.agents.initialize_agent", "langchain.llms.OpenAI", "langchain.agents.Tool", "langchain.chat_models.ChatOpenAI" ]
[((786, 807), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (792, 807), False, 'from langchain.llms import OpenAI, Cohere, HuggingFaceHub\n'), ((815, 840), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (825, 840), False, 'fro...
import langchain from dotenv import load_dotenv from langchain_openai import ChatOpenAI, OpenAI from langchain.schema import HumanMessage, AIMessage, SystemMessage from langchain.prompts import PromptTemplate, FewShotPromptTemplate from langchain.output_parsers import CommaSeparatedListOutputParser from langchain.cache...
[ "langchain_openai.ChatOpenAI", "langchain.cache.InMemoryCache", "langchain.prompts.PromptTemplate" ]
[((423, 436), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (434, 436), False, 'from dotenv import load_dotenv\n'), ((459, 474), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (472, 474), False, 'from langchain.cache import InMemoryCache\n'), ((508, 541), 'langchain_openai.ChatOpenAI', 'Ch...
import json from pathlib import Path from typing import Dict, List import langchain import numpy as np import typer from langchain.cache import SQLiteCache from langchain.llms import OpenAI from tqdm import tqdm langchain.llm_cache = SQLiteCache(database_path=".langchain.db") def _is_daster_empl(title: str) -> bool...
[ "langchain.llms.OpenAI", "langchain.cache.SQLiteCache" ]
[((236, 278), 'langchain.cache.SQLiteCache', 'SQLiteCache', ([], {'database_path': '""".langchain.db"""'}), "(database_path='.langchain.db')\n", (247, 278), False, 'from langchain.cache import SQLiteCache\n'), ((905, 952), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0.1)', 'model_name': '"""gpt-4-32k"""'}...
import langchain.vectorstores.opensearch_vector_search as ovs from opensearchpy import OpenSearch, RequestsHttpConnection, AWSV4SignerAuth, helpers from langchain.vectorstores import OpenSearchVectorSearch def create_ovs_client( collection_id, index_name, region, boto3_session, bedrock_embeddings...
[ "langchain.vectorstores.OpenSearchVectorSearch" ]
[((470, 515), 'opensearchpy.AWSV4SignerAuth', 'AWSV4SignerAuth', (['credentials', 'region', 'service'], {}), '(credentials, region, service)\n', (485, 515), False, 'from opensearchpy import OpenSearch, RequestsHttpConnection, AWSV4SignerAuth, helpers\n'), ((543, 724), 'opensearchpy.OpenSearch', 'OpenSearch', ([], {'hos...
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by redu...
[ "langchain.load.load.loads", "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps" ]
[((1586, 1613), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1603, 1613), False, 'import logging\n'), ((5793, 5811), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (5809, 5811), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), (...
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks....
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages....
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (103...
from langchain.chains.router import MultiPromptChain from langchain.chains.router.llm_router import LLMRouterChain, RouterOutputParser from langchain.prompts import PromptTemplate from langchain.chat_models import ChatOpenAI from langchain.prompts import ChatPromptTemplate from langchain.chains import LLMChain from ap...
[ "langchain.chains.router.llm_router.LLMRouterChain.from_llm", "langchain.chat_models.ChatOpenAI", "langchain.chains.router.MultiPromptChain", "langchain.chains.router.llm_router.RouterOutputParser", "langchain.chains.LLMChain", "langchain.prompts.ChatPromptTemplate.from_template" ]
[((3977, 4033), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': 'ChatGPTModel.GPT3.value'}), '(temperature=0, model=ChatGPTModel.GPT3.value)\n', (3987, 4033), False, 'from langchain.chat_models import ChatOpenAI\n'), ((4531, 4574), 'langchain.prompts.ChatPromptTemplate.from_templa...
from __future__ import annotations import asyncio import functools import logging import os import warnings from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( Any, AsyncGenerator, Dict, Generator, List, Optional, Type, Type...
[ "langchain.schema.get_buffer_string", "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainT...
[((1114, 1141), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1131, 1141), False, 'import logging\n'), ((1286, 1329), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1296, 1329), False, 'from contextvars i...
import argparse import json import logging import os import pathlib from typing import Dict, List, Union, Optional import langchain import pandas as pd import tiktoken import wandb from langchain import LLMChain, FAISS from langchain.cache import SQLiteCache from langchain.chains import HypotheticalDocumentEmbedder fr...
[ "langchain.docstore.document.Document", "langchain.text_splitter.MarkdownTextSplitter", "langchain.cache.SQLiteCache", "langchain.chat_models.ChatOpenAI", "langchain.document_loaders.NotebookLoader", "langchain.text_splitter.PythonCodeTextSplitter", "langchain.text_splitter.TokenTextSplitter", "langch...
[((902, 943), 'langchain.cache.SQLiteCache', 'SQLiteCache', ([], {'database_path': '"""langchain.db"""'}), "(database_path='langchain.db')\n", (913, 943), False, 'from langchain.cache import SQLiteCache\n'), ((954, 981), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (971, 981), False, 'i...
import langchain from dotenv import load_dotenv from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from rmrkl import ChatZeroShotAgent, RetryAgentExecutor from .prompt import FORMAT_INSTRUCTIONS, QUESTION_PROMPT, SUFFIX from .tools import make_tools, Doc, Text,search_texts, load_texts imp...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((329, 342), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (340, 342), False, 'from dotenv import load_dotenv\n'), ((2064, 2077), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (2074, 2077), False, 'import time\n'), ((1632, 1784), 'rmrkl.ChatZeroShotAgent.from_llm_and_tools', 'ChatZeroShotAgent.from_llm_...
import os import json import time from typing import List import faiss import pypdf import random import itertools import text_utils import pandas as pd import altair as alt import streamlit as st from io import StringIO from llama_index import Document from langchain.llms import Anthropic from langchain.chains import ...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.retrievers.SVMRetriever.from_texts", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.chains.RetrievalQA.from_chain_type", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chat_models.ChatOpenAI", "langchain.vectorst...
[((13312, 13350), 'streamlit.sidebar.image', 'st.sidebar.image', (['"""img/diagnostic.jpg"""'], {}), "('img/diagnostic.jpg')\n", (13328, 13350), True, 'import streamlit as st\n'), ((15130, 15159), 'streamlit.header', 'st.header', (['"""`Auto-evaluator`"""'], {}), "('`Auto-evaluator`')\n", (15139, 15159), True, 'import ...
# general imports from constants import * # streamlit imports import streamlit as st from utils import * from streamlit_lottie import st_lottie # llama index imports import openai from llama_index import ( VectorStoreIndex, download_loader, ServiceContext, set_global_service_context, ) from llama_inde...
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((1017, 1080), 'llama_index.llms.OpenAI', 'OpenAI', ([], {'model': '"""gpt-4-1106-preview"""', 'system_prompt': 'system_prompt'}), "(model='gpt-4-1106-preview', system_prompt=system_prompt)\n", (1023, 1080), False, 'from llama_index.llms import OpenAI\n'), ((1187, 1248), 'llama_index.ServiceContext.from_defaults', 'Se...
#%% import pandas as pd from utils import get_random_string from dotenv import load_dotenv import os import langchain from langchain.prompts import PromptTemplate from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from openai import OpenAI import json import requests import datetime import...
[ "langchain.prompts.PromptTemplate.from_template", "langchain.chat_models.ChatOpenAI" ]
[((347, 360), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (358, 360), False, 'from dotenv import load_dotenv\n'), ((368, 416), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (378, 416), False, 'from l...
import os #from dotenv import load_dotenv import openai import langchain os.environ["OPENAI_API_KEY"] ="" os.environ["SQL_SERVER_USERNAME"] = "" os.environ["SQL_SERVER_ENDPOINT"] = "" os.environ["SQL_SERVER_PASSWORD"] = "" os.environ["SQL_SERVER_DATABASE"] = "" from sqlalchemy import create_engine from sqlalchemy....
[ "langchain.agents.create_sql_agent", "langchain.agents.agent_toolkits.SQLDatabaseToolkit", "langchain.llms.OpenAI", "langchain.sql_database.SQLDatabase.from_uri" ]
[((785, 808), 'sqlalchemy.engine.url.URL.create', 'URL.create', ([], {}), '(**db_config)\n', (795, 808), False, 'from sqlalchemy.engine.url import URL\n'), ((814, 842), 'langchain.sql_database.SQLDatabase.from_uri', 'SQLDatabase.from_uri', (['db_url'], {}), '(db_url)\n', (834, 842), False, 'from langchain.sql_database ...
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging import threading import weakref from concurrent.futures import Future, ThreadPoolExecutor, wait from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast from uuid import UUID import langsmith f...
[ "langchain.callbacks.tracers.langchain._get_executor", "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((672, 699), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (689, 699), False, 'import logging\n'), ((755, 772), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (770, 772), False, 'import weakref\n'), ((3430, 3447), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (3445, 3...
import os import re from uuid import UUID from typing import Any, Dict, List, Optional, Union import asyncio import langchain import streamlit as st from langchain.schema import LLMResult from langchain.chat_models import ChatOpenAI from langchain.agents import Tool from langchain.agents import AgentType from langcha...
[ "langchain.agents.initialize_agent", "langchain.memory.ConversationBufferMemory", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.agents.Tool" ]
[((815, 826), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (824, 826), False, 'import os\n'), ((6031, 6120), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)', 'openai_api_key': 'openai_api_key'}), "(model_name='gpt-3.5-turbo', temperature=0, openai_api_key...
from approaches.index.store.cosmos_index_store import CosmosIndexStore from llama_index import StorageContext from approaches.index.store.cosmos_doc_store import CosmosDocumentStore from llama_index import load_index_from_storage import os import openai from langchain.chat_models import AzureChatOpenAI from langchain....
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((832, 845), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (843, 845), False, 'from dotenv import load_dotenv\n'), ((1039, 1074), 'os.environ.get', 'os.environ.get', (['"""AZURE_OPENAI_BASE"""'], {}), "('AZURE_OPENAI_BASE')\n", (1053, 1074), False, 'import os\n'), ((1098, 1153), 'os.environ.get', 'os.environ....
from abc import ABC, abstractmethod from typing import List, Optional from pydantic import BaseModel, Extra, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import ( AIMessage, BaseLanguageMod...
[ "langchain.schema.ChatResult", "langchain.schema.ChatGeneration", "langchain.schema.HumanMessage", "langchain.schema.AIMessage", "langchain.schema.LLMResult", "langchain.callbacks.get_callback_manager" ]
[((568, 605), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (573, 605), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((696, 739), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager'}), '(default_factory=get_ca...
import logging import os import pprint import uuid from typing import List import chromadb import gradio as gr import requests import zhipuai from bs4 import BeautifulSoup from dotenv import load_dotenv, find_dotenv # Import langchain stuff from langchain.chains import ConversationalRetrievalChain from langchain.docum...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain_community.vectorstores.chroma.Chroma.from_documents", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain_community.vectorstores.chroma.Chroma", "langchain_core.prompts.PromptTemplate"...
[((1392, 1490), 'llms.zhipuai_llm.ZhipuAILLM', 'ZhipuAILLM', ([], {'model': '"""chatglm_turbo"""', 'temperature': '(0.9)', 'top_p': '(0.1)', 'zhipuai_api_key': 'zhipuai.api_key'}), "(model='chatglm_turbo', temperature=0.9, top_p=0.1,\n zhipuai_api_key=zhipuai.api_key)\n", (1402, 1490), False, 'from llms.zhipuai_llm ...
"""An example of how to test Python code generating prompts""" import re # Brining some "prompt generator" classes from promptimize.prompt_cases import LangchainPromptCase # Bringing some useful eval function that help evaluating and scoring responses # eval functions have a handle on the prompt object and are expect...
[ "langchain.output_parsers.ResponseSchema", "langchain.output_parsers.StructuredOutputParser.from_response_schemas", "langchain.PromptTemplate" ]
[((1146, 1208), 'langchain.output_parsers.StructuredOutputParser.from_response_schemas', 'StructuredOutputParser.from_response_schemas', (['response_schemas'], {}), '(response_schemas)\n', (1190, 1208), False, 'from langchain.output_parsers import StructuredOutputParser, ResponseSchema\n'), ((2218, 2382), 'langchain.Pr...
""" The ``mlflow.langchain`` module provides an API for logging and loading LangChain models. This module exports multivariate LangChain models in the langchain flavor and univariate LangChain models in the pyfunc flavor: LangChain (native) format This is the main flavor that can be accessed with LangChain APIs. :...
[ "langchain.agents.initialize_agent", "langchain.chains.loading.load_chain" ]
[((2012, 2046), 'logging.getLogger', 'logging.getLogger', (['mlflow.__name__'], {}), '(mlflow.__name__)\n', (2029, 2046), False, 'import logging\n'), ((11731, 11807), 'mlflow.utils.environment._validate_env_arguments', '_validate_env_arguments', (['conda_env', 'pip_requirements', 'extra_pip_requirements'], {}), '(conda...
# Import the necessary libraries import random import time from llama_index.llms import OpenAI import streamlit as st from llama_index import VectorStoreIndex, ServiceContext, StorageContext, set_global_service_context from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_index.embeddings import...
[ "langchain_openai.ChatOpenAI", "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((855, 895), 'streamlit.title', 'st.title', (['"""🦜🔗 Tourism Assistant Chatbot"""'], {}), "('🦜🔗 Tourism Assistant Chatbot')\n", (863, 895), True, 'import streamlit as st\n'), ((5721, 5781), 'llama_index.set_global_service_context', 'set_global_service_context', (['st.session_state.service_context'], {}), '(st.sess...
# This code sets up the necessary components, interacts with the LangChain tool and ChatOpenAI model to perform text summarization, # and provides a user interface for input and output. from langchain.document_loaders import UnstructuredFileLoader # Importing necessary modules from langchain.document_loaders import ...
[ "langchain.chains.summarize.load_summarize_chain", "langchain.document_loaders.UnstructuredFileLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chat_models.ChatOpenAI", "langchain.document_loaders.UnstructuredPDFLoader", "langchain.prompts.PromptTemplate" ]
[((5769, 5891), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Positive summarizer"""', 'page_icon': '"""📖"""', 'layout': '"""wide"""', 'initial_sidebar_state': '"""collapsed"""'}), "(page_title='Positive summarizer', page_icon='📖', layout=\n 'wide', initial_sidebar_state='collapsed')\n...
import streamlit as st from streamlit_chat import message import pandas as pd from langchain.llms import OpenAI import os from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationSummaryBufferMemory import plotly.express from streamlit_searchbox import st_searchbox from typing import List, ...
[ "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1329, 1342), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (1340, 1342), False, 'from dotenv import load_dotenv\n'), ((1378, 1486), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""PubMeta.ai"""', 'page_icon': '"""⚕️"""', 'layout': '"""wide"""', 'initial_sidebar_state': '"""auto"""...
from typing import Any, Dict, List, Optional from langchain import PromptTemplate ,LLMChain import langchain from langchain.chat_models import ChatOpenAI ,AzureChatOpenAI from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler import sys import re import argparse import os print(sys.path) sys.p...
[ "langchain.LLMChain", "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.prompts.chat.ChatPromptTemplate", "langchain.schema.SystemMessage", "langchain.prompts.chat.HumanMessagePromptTemplate.from_template", "langchain.PromptTemplate" ]
[((315, 335), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (330, 335), False, 'import sys\n'), ((3893, 3960), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_template', 'input_variables': "['essay']"}), "(template=prompt_template, input_variables=['essay'])\n", (3907, 3960...
import streamlit as st from dotenv import load_dotenv load_dotenv() import os import tempfile from llama_index import SimpleDirectoryReader, StorageContext, LLMPredictor from llama_index import VectorStoreIndex from llama_index import ServiceContext from llama_index.embeddings.langchain import LangchainEmbedding from...
[ "langchain.embeddings.CohereEmbeddings", "langchain.chat_models.ChatOpenAI" ]
[((55, 68), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (66, 68), False, 'from dotenv import load_dotenv\n'), ((860, 890), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {}), '()\n', (888, 890), False, 'from llama_index import SimpleDirectoryReader, StorageContext, LLMPredic...
import os import langchain from langchain.utilities import SerpAPIWrapper from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType from langchain.chat_models import ChatOpenAI os.environ['OPENAI_API_KEY'] = "" os.environ['SERPAPI_API_KEY'] = "" llm = ChatOpenAI(temperature=0, model=...
[ "langchain.agents.initialize_agent", "langchain.utilities.SerpAPIWrapper", "langchain.agents.Tool", "langchain.chat_models.ChatOpenAI" ]
[((288, 341), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': '"""gpt-3.5-turbo-0613"""'}), "(temperature=0, model='gpt-3.5-turbo-0613')\n", (298, 341), False, 'from langchain.chat_models import ChatOpenAI\n'), ((352, 368), 'langchain.utilities.SerpAPIWrapper', 'SerpAPIWrapper', (...
import streamlit as st from dotenv import load_dotenv import os from htmlTemplates import css, bot_template, user_template import langchain from langchain.document_loaders import GitLoader from langchain.text_splitter import ( RecursiveCharacterTextSplitter, Language, ) from langchain.text_splitter import Recur...
[ "langchain.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.memory.ConversationBufferMemory", "langchain.document_loaders.GitLoader", "langchain.vectorstores.DeepLake", "langchain.chat_models.ChatOpenAI", "langchain.text_splitter.RecursiveCharacterTextSplitter.from_language", "langcha...
[((822, 895), 'langchain.document_loaders.GitLoader', 'GitLoader', ([], {'clone_url': 'github_url', 'repo_path': 'local_path', 'branch': 'repo_branch'}), '(clone_url=github_url, repo_path=local_path, branch=repo_branch)\n', (831, 895), False, 'from langchain.document_loaders import GitLoader\n'), ((2876, 2915), 'langch...
import json from llama_index.core.service_context_elements.llm_predictor import LLMPredictor from llama_index.core.utilities.sql_wrapper import SQLDatabase from llama_index.core.response_synthesizers import get_response_synthesizer from llama_index.embeddings.langchain import LangchainEmbedding from llama_index.core.re...
[ "langchain.agents.initialize_agent", "langchain_community.chat_models.ChatOpenAI" ]
[((3043, 3075), 'app.database.dbc.get_llm_by_name', 'dbc.get_llm_by_name', (['db', 'llmName'], {}), '(db, llmName)\n', (3062, 3075), False, 'from app.database import dbc\n'), ((4079, 4112), 'app.database.dbc.get_project_by_name', 'dbc.get_project_by_name', (['db', 'name'], {}), '(db, name)\n', (4102, 4112), False, 'fro...
# This is an example of integrating a LLM with streamlit import streamlit as st import os import openai import langchain from langchain.llms import OpenAI from langchain import PromptTemplate #from dotenv import load_dotenv # Specify the path to the .env file #dotenv_path = os.path.join(os.path.dirname(__file__), '.en...
[ "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((390, 459), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Globalize Email"""', 'page_icon': '""":robot:"""'}), "(page_title='Globalize Email', page_icon=':robot:')\n", (408, 459), True, 'import streamlit as st\n'), ((462, 489), 'streamlit.header', 'st.header', (['"""Globalize Text"""'], {...
import sys sys.stdout.reconfigure(encoding="utf-8") sys.stdin.reconfigure(encoding="utf-8") import streamlit as st import streamlit.components.v1 as components import re import random CODE_BUILD_KG = """ # 准备 GraphStore os.environ['NEBULA_USER'] = "root" os.environ['NEBULA_PASSWORD'] = "nebula" # default passwor...
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((12, 52), 'sys.stdout.reconfigure', 'sys.stdout.reconfigure', ([], {'encoding': '"""utf-8"""'}), "(encoding='utf-8')\n", (34, 52), False, 'import sys\n'), ((53, 92), 'sys.stdin.reconfigure', 'sys.stdin.reconfigure', ([], {'encoding': '"""utf-8"""'}), "(encoding='utf-8')\n", (74, 92), False, 'import sys\n'), ((2988, 3...
from langchain.agents import AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.tools import Tool, StructuredTool from langchain.prompts import StringPromptTemplate from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.llms import VertexAI from typing imp...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.schema.AgentAction", "langchain.llms.VertexAI", "langchain.schema.AgentFinish", "langchain.callbacks.FileCallbackHandler", "langchain.chains.LLMChain" ]
[((1046, 1098), 'os.makedirs', 'os.makedirs', (['f"""./results/{timestamp}"""'], {'exist_ok': '(True)'}), "(f'./results/{timestamp}', exist_ok=True)\n", (1057, 1098), False, 'import os\n'), ((1332, 1364), 'logging.getLogger', 'logging.getLogger', (['"""info_logger"""'], {}), "('info_logger')\n", (1349, 1364), False, 'i...
import langchain from dotenv import load_dotenv from langchain.chains import FlareChain from langchain.chat_models import ChatOpenAI from langchain.embeddings import OpenAIEmbeddings from langchain.llms import OpenAI from langchain.vectorstores import FAISS langchain.verbose = True load_dotenv() # FAISSで保存されたベクトルを読み...
[ "langchain.vectorstores.FAISS.load_local", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.chains.FlareChain.from_llm", "langchain.embeddings.OpenAIEmbeddings" ]
[((285, 298), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (296, 298), False, 'from dotenv import load_dotenv\n'), ((336, 354), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (352, 354), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((360, 403), 'langchain.vect...
import os from dotenv import load_dotenv import streamlit as st from langchain.chains import LLMChain from langchain import PromptTemplate from genai.credentials import Credentials from genai.extensions.langchain import LangChainInterface from genai.schemas import GenerateParams load_dotenv() api_key = os.getenv("GE...
[ "langchain.chains.LLMChain", "langchain.PromptTemplate" ]
[((283, 296), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (294, 296), False, 'from dotenv import load_dotenv\n'), ((307, 335), 'os.getenv', 'os.getenv', (['"""GENAI_KEY"""', 'None'], {}), "('GENAI_KEY', None)\n", (316, 335), False, 'import os\n'), ((351, 379), 'os.getenv', 'os.getenv', (['"""GENAI_API"""', '...
import os from datetime import datetime, timezone from dotenv import load_dotenv from langchain.agents import AgentType, initialize_agent from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationBufferMemory from langchain.prompts import MessagesPlaceholder from src.xm_group_tools import ...
[ "langchain.agents.initialize_agent", "langchain.memory.ConversationBufferMemory", "langchain.prompts.MessagesPlaceholder", "langchain.chat_models.ChatOpenAI" ]
[((458, 471), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (469, 471), False, 'from dotenv import load_dotenv\n'), ((482, 509), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (491, 509), False, 'import os\n'), ((742, 770), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([...
"""Push and pull to the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain.load.dump import dumps from langchain.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Opti...
[ "langchain.load.load.loads", "langchainhub.Client", "langchain.load.dump.dumps" ]
[((671, 703), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (677, 703), False, 'from langchainhub import Client\n'), ((1886, 1899), 'langchain.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1891, 1899), False, 'from langchain.load.dump import dumps\...
"""Base interface that all chains should implement.""" import inspect import json import logging import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import Field, root_validator, validator import langchain from lang...
[ "langchain.schema.RunInfo", "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.load.dump.dumpd", "langchain.callbacks.manager.CallbackManager.configure" ]
[((702, 729), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (719, 729), False, 'import logging\n'), ((2435, 2468), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2440, 2468), False, 'from pydantic import Field, root_validator...
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import functools import inspect import logging import uuid from enum import Enum from typing import ( TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, ...
[ "langchain.schema.messages.messages_from_dict", "langchain._api.warn_deprecated", "langchain.schema.runnable.config.get_executor_for_config", "langchain.evaluation.schema.EvaluatorType", "langchain.smith.evaluation.name_generation.random_name", "langchain.smith.evaluation.StringRunEvaluatorChain.from_run_...
[((1724, 1751), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1741, 1751), False, 'import logging\n'), ((33983, 34008), 'langchain.callbacks.tracers.evaluation.wait_for_all_evaluators', 'wait_for_all_evaluators', ([], {}), '()\n', (34006, 34008), False, 'from langchain.callbacks.tracers...
from pydantic import BaseModel, Field import os from langchain import OpenAI from langchain.chat_models import ChatOpenAI from langchain.agents import initialize_agent, Tool # from langchain.chains import PALChain from langchain.chains.conversation.memory import ConversationBufferMemory from langchain import Pr...
[ "langchain.agents.AgentExecutor", "langchain.memory.ConversationBufferMemory", "langchain.prompts.MessagesPlaceholder", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.FAISS.from_documents", "langchain.agents.agent_toolkits.create_retriever_tool", "langchain.schema.messages.SystemMessage", ...
[((3412, 3472), 'streamlit.markdown', 'st.markdown', (['hide_share_button_style'], {'unsafe_allow_html': '(True)'}), '(hide_share_button_style, unsafe_allow_html=True)\n', (3423, 3472), True, 'import streamlit as st\n'), ((3474, 3537), 'streamlit.markdown', 'st.markdown', (['hide_star_and_github_style'], {'unsafe_allow...
import streamlit as st import openai import langchain # from langchain import PromptTemplate, LLMChain # from langchain.llms import OpenAI # # Set your OpenAI API key # openai_api_key = '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' # # Create a sidebar for language selection # st.sidebar.title('Translation A...
[ "langchain.LLMChain", "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((3099, 3134), 'streamlit.sidebar.title', 'st.sidebar.title', (['"""Translation App"""'], {}), "('Translation App')\n", (3115, 3134), True, 'import streamlit as st\n'), ((3216, 3265), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""Input Language"""', 'languages'], {}), "('Input Language', languages)\n", ...
import langchain.llms from langchain import GoogleSearchAPIWrapper, LLMChain from langchain.agents import initialize_agent, AgentType, Tool, ZeroShotAgent, AgentExecutor from langchain.schema import BaseMemory def setup_agent(llm: langchain.llms.BaseLLM, memory: BaseMemory): search = GoogleSearchAPIWrapper() ...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.LLMChain", "langchain.agents.ZeroShotAgent.create_prompt", "langchain.agents.ZeroShotAgent", "langchain.GoogleSearchAPIWrapper", "langchain.agents.Tool" ]
[((291, 315), 'langchain.GoogleSearchAPIWrapper', 'GoogleSearchAPIWrapper', ([], {}), '()\n', (313, 315), False, 'from langchain import GoogleSearchAPIWrapper, LLMChain\n'), ((833, 948), 'langchain.agents.ZeroShotAgent.create_prompt', 'ZeroShotAgent.create_prompt', (['tools'], {'prefix': 'prefix', 'suffix': 'suffix', '...
import itertools from langchain.cache import InMemoryCache, SQLiteCache import langchain import pandas as pd from certa.utils import merge_sources from certa.explain import CertaExplainer from datetime import datetime import os import ellmer.models import ellmer.metrics from time import sleep, time import json import t...
[ "langchain.cache.InMemoryCache", "langchain.cache.SQLiteCache" ]
[((8572, 8636), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Run saliency experiments."""'}), "(description='Run saliency experiments.')\n", (8595, 8636), False, 'import argparse\n'), ((598, 613), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (611, 613), False, 'f...
# Databricks notebook source # MAGIC %md # MAGIC # 3. Chatbotの作成とデプロイ # MAGIC # MAGIC <br/> # MAGIC <img src="https://github.com/naoyaabe-db/public_demo_images/blob/3380b6d73937cd95efae845799c37de910b7394c/rag_demo_images/diagram_notebook3.png?raw=true" style="float: right" width="1000px"> # MAGIC <br/> # MAGIC # MAGIC...
[ "langchain.chains.RetrievalQA.from_chain_type", "langchain.vectorstores.DatabricksVectorSearch", "langchain.prompts.PromptTemplate", "langchain.chat_models.ChatDatabricks" ]
[((5946, 5996), 'mlflow.deployments.get_deploy_client', 'mlflow.deployments.get_deploy_client', (['"""databricks"""'], {}), "('databricks')\n", (5982, 5996), False, 'import mlflow\n'), ((6909, 6974), 'langchain.chat_models.ChatDatabricks', 'ChatDatabricks', ([], {'endpoint': 'chat_model_endpoint_name', 'max_tokens': '(...
# Import Langchain modules from langchain.document_loaders import PyPDFLoader from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import FAISS from langchain.chains import RetrievalQA from langchain.llms import OpenAI # Impo...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.vectorstores.FAISS.from_documents", "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.OpenAIEmbeddings" ]
[((573, 606), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (596, 606), False, 'import warnings\n'), ((712, 808), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)s - %(levelname)s - %(message)s"""'}), "(level=logging.IN...
import streamlit as st import dotenv import langchain import json from cassandra.cluster import Session from cassandra.query import PreparedStatement from langchain.agents.agent_toolkits import create_retriever_tool, create_conversational_retrieval_agent from langchain.chat_models import ChatOpenAI from langchain.emb...
[ "langchain.chat_models.ChatOpenAI", "langchain.schema.Document", "langchain.agents.agent_toolkits.create_conversational_retrieval_agent", "langchain.agents.agent_toolkits.create_retriever_tool", "langchain.schema.SystemMessage", "langchain.embeddings.OpenAIEmbeddings" ]
[((5021, 5054), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (5039, 5054), True, 'import streamlit as st\n'), ((5462, 5502), 'streamlit.chat_input', 'st.chat_input', ([], {'placeholder': '"""Ask chatbot"""'}), "(placeholder='Ask chatbot')\n", (5475, 5502), True...
import os import langchain import streamlit as st from dotenv import load_dotenv from langchain.chat_models import ChatOpenAI from langchain.cache import InMemoryCache from langchain.prompts import SystemMessagePromptTemplate, HumanMessagePromptTemplate, AIMessagePromptTemplate, ChatPromptTemplate, PromptTemplate # C...
[ "langchain.cache.InMemoryCache", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.prompts.SystemMessagePromptTemplate.from_template" ]
[((326, 345), 'dotenv.load_dotenv', 'load_dotenv', (['""".env"""'], {}), "('.env')\n", (337, 345), False, 'from dotenv import load_dotenv\n'), ((356, 377), 'os.environ.get', 'os.environ.get', (['"""key"""'], {}), "('key')\n", (370, 377), False, 'import os\n'), ((416, 431), 'langchain.cache.InMemoryCache', 'InMemoryCach...
import os import dotenv dotenv.load_dotenv() ### Load the credentials api_key = os.getenv("API_KEY", None) ibm_cloud_url = os.getenv("IBM_CLOUD_URL", None) project_id = os.getenv("PROJECT_ID", None) HUGGINGFACEHUB_API_TOKEN = os.getenv("HUGGINGFACEHUB_API_TOKEN", None) min_new_tokens=1 max_new_tokens=300 temperature...
[ "langchain.embeddings.HuggingFaceHubEmbeddings" ]
[((24, 44), 'dotenv.load_dotenv', 'dotenv.load_dotenv', ([], {}), '()\n', (42, 44), False, 'import dotenv\n'), ((81, 107), 'os.getenv', 'os.getenv', (['"""API_KEY"""', 'None'], {}), "('API_KEY', None)\n", (90, 107), False, 'import os\n'), ((124, 156), 'os.getenv', 'os.getenv', (['"""IBM_CLOUD_URL"""', 'None'], {}), "('...
from typing import Any, Dict, List, Optional from .few_shot_agent import FewShotAgent from .few_shot_agent import FewShotAgentExecutor from langchain import LLMChain from langchain.tools.base import BaseTool from typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, Union from .prompts import * import nes...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((2122, 2142), 'nest_asyncio.apply', 'nest_asyncio.apply', ([], {}), '()\n', (2140, 2142), False, 'import nest_asyncio\n'), ((772, 804), 'langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler', 'StreamingStdOutCallbackHandler', ([], {}), '()\n', (802, 804), False, 'from langchain.callbacks.streaming_stdo...
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator impor...
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMRes...
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, ...
from fastapi import FastAPI from langchain import ConversationChain from langchain.chat_models import ChatOpenAI from scripts.utils import MEMORY from scripts.doc_loader import load_document from lanarky import LangchainRouter from starlette.requests import Request from starlette.templating import Jinja2Templates from...
[ "langchain.chat_models.ChatOpenAI" ]
[((352, 369), 'config.set_environment', 'set_environment', ([], {}), '()\n', (367, 369), False, 'from config import set_environment\n'), ((377, 386), 'fastapi.FastAPI', 'FastAPI', ([], {}), '()\n', (384, 386), False, 'from fastapi import FastAPI\n'), ((570, 619), 'starlette.templating.Jinja2Templates', 'Jinja2Templates...
# %% import torch import os from llama_index.core import VectorStoreIndex, SimpleDirectoryReader from llama_index.llms.huggingface import HuggingFaceInferenceAPI from llama_index.core import Settings from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_index.embeddings.langchain import Langch...
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((1018, 1239), 'llama_index.core.PromptTemplate', 'PromptTemplate', (['"""Your job is to summarize different sections of the document given to you.Write a response that appropriately completes the request given to you.\n\n### Instruction:\n{query_str}\n\n### Response:"""'], {}), '(\n """Your job is to summarize dif...
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, ...
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n'...
import streamlit as st from langchain import PromptTemplate from utils.studio_style import apply_studio_style from utils.studio_style import keyword_label, sentiment_label from utils import langchain from utils import bedrock from utils import config from datetime import datetime import pandas as pd import json import ...
[ "langchain.PromptTemplate" ]
[((329, 402), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Summarize Product Reviews"""', 'page_icon': '"""🛒"""'}), "(page_title='Summarize Product Reviews', page_icon='🛒')\n", (347, 402), True, 'import streamlit as st\n'), ((415, 438), 'utils.config.get_background', 'config.get_backgrou...
docs = """When and under what conditions can I apply to your graduate programs? Graduate student admissions are made in the fall and spring semesters specified in the academic calendar. Minimum application requirements: To have the undergraduate degree required in the program application requirements To have at least ...
[ "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.cache.SQLiteCache" ]
[((2056, 2075), 'dotenv.load_dotenv', 'load_dotenv', (['""".env"""'], {}), "('.env')\n", (2067, 2075), False, 'from dotenv import load_dotenv\n'), ((2086, 2118), 'os.environ.get', 'os.environ.get', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (2100, 2118), False, 'import os\n'), ((2200, 2242), 'langchain.cac...