Spaces:
No application file
No application file
| title: ❓ FAQs | |
| description: 'Collections of all the frequently asked questions' | |
| <AccordionGroup> | |
| <Accordion title="Does Embedchain support OpenAI's Assistant APIs?"> | |
| Yes, it does. Please refer to the [OpenAI Assistant docs page](/examples/openai-assistant). | |
| </Accordion> | |
| <Accordion title="How to use MistralAI language model?"> | |
| Use the model provided on huggingface: `mistralai/Mistral-7B-v0.1` | |
| <CodeGroup> | |
| ```python main.py | |
| import os | |
| from embedchain import App | |
| os.environ["HUGGINGFACE_ACCESS_TOKEN"] = "hf_your_token" | |
| app = App.from_config("huggingface.yaml") | |
| ``` | |
| ```yaml huggingface.yaml | |
| llm: | |
| provider: huggingface | |
| config: | |
| model: 'mistralai/Mistral-7B-v0.1' | |
| temperature: 0.5 | |
| max_tokens: 1000 | |
| top_p: 0.5 | |
| stream: false | |
| embedder: | |
| provider: huggingface | |
| config: | |
| model: 'sentence-transformers/all-mpnet-base-v2' | |
| ``` | |
| </CodeGroup> | |
| </Accordion> | |
| <Accordion title="How to use ChatGPT 4 turbo model released on OpenAI DevDay?"> | |
| Use the model `gpt-4-turbo` provided my openai. | |
| <CodeGroup> | |
| ```python main.py | |
| import os | |
| from embedchain import App | |
| os.environ['OPENAI_API_KEY'] = 'xxx' | |
| # load llm configuration from gpt4_turbo.yaml file | |
| app = App.from_config(config_path="gpt4_turbo.yaml") | |
| ``` | |
| ```yaml gpt4_turbo.yaml | |
| llm: | |
| provider: openai | |
| config: | |
| model: 'gpt-4-turbo' | |
| temperature: 0.5 | |
| max_tokens: 1000 | |
| top_p: 1 | |
| stream: false | |
| ``` | |
| </CodeGroup> | |
| </Accordion> | |
| <Accordion title="How to use GPT-4 as the LLM model?"> | |
| <CodeGroup> | |
| ```python main.py | |
| import os | |
| from embedchain import App | |
| os.environ['OPENAI_API_KEY'] = 'xxx' | |
| # load llm configuration from gpt4.yaml file | |
| app = App.from_config(config_path="gpt4.yaml") | |
| ``` | |
| ```yaml gpt4.yaml | |
| llm: | |
| provider: openai | |
| config: | |
| model: 'gpt-4' | |
| temperature: 0.5 | |
| max_tokens: 1000 | |
| top_p: 1 | |
| stream: false | |
| ``` | |
| </CodeGroup> | |
| </Accordion> | |
| <Accordion title="I don't have OpenAI credits. How can I use some open source model?"> | |
| <CodeGroup> | |
| ```python main.py | |
| from embedchain import App | |
| # load llm configuration from opensource.yaml file | |
| app = App.from_config(config_path="opensource.yaml") | |
| ``` | |
| ```yaml opensource.yaml | |
| llm: | |
| provider: gpt4all | |
| config: | |
| model: 'orca-mini-3b-gguf2-q4_0.gguf' | |
| temperature: 0.5 | |
| max_tokens: 1000 | |
| top_p: 1 | |
| stream: false | |
| embedder: | |
| provider: gpt4all | |
| config: | |
| model: 'all-MiniLM-L6-v2' | |
| ``` | |
| </CodeGroup> | |
| </Accordion> | |
| <Accordion title="How to stream response while using OpenAI model in Embedchain?"> | |
| You can achieve this by setting `stream` to `true` in the config file. | |
| <CodeGroup> | |
| ```yaml openai.yaml | |
| llm: | |
| provider: openai | |
| config: | |
| model: 'gpt-3.5-turbo' | |
| temperature: 0.5 | |
| max_tokens: 1000 | |
| top_p: 1 | |
| stream: true | |
| ``` | |
| ```python main.py | |
| import os | |
| from embedchain import App | |
| os.environ['OPENAI_API_KEY'] = 'sk-xxx' | |
| app = App.from_config(config_path="openai.yaml") | |
| app.add("https://www.forbes.com/profile/elon-musk") | |
| response = app.query("What is the net worth of Elon Musk?") | |
| # response will be streamed in stdout as it is generated. | |
| ``` | |
| </CodeGroup> | |
| </Accordion> | |
| <Accordion title="How to persist data across multiple app sessions?"> | |
| Set up the app by adding an `id` in the config file. This keeps the data for future use. You can include this `id` in the yaml config or input it directly in `config` dict. | |
| ```python app1.py | |
| import os | |
| from embedchain import App | |
| os.environ['OPENAI_API_KEY'] = 'sk-xxx' | |
| app1 = App.from_config(config={ | |
| "app": { | |
| "config": { | |
| "id": "your-app-id", | |
| } | |
| } | |
| }) | |
| app1.add("https://www.forbes.com/profile/elon-musk") | |
| response = app1.query("What is the net worth of Elon Musk?") | |
| ``` | |
| ```python app2.py | |
| import os | |
| from embedchain import App | |
| os.environ['OPENAI_API_KEY'] = 'sk-xxx' | |
| app2 = App.from_config(config={ | |
| "app": { | |
| "config": { | |
| # this will persist and load data from app1 session | |
| "id": "your-app-id", | |
| } | |
| } | |
| }) | |
| response = app2.query("What is the net worth of Elon Musk?") | |
| ``` | |
| </Accordion> | |
| </AccordionGroup> | |
| #### Still have questions? | |
| If docs aren't sufficient, please feel free to reach out to us using one of the following methods: | |
| <Snippet file="get-help.mdx" /> | |