Spaces:
Runtime error
Runtime error
| title: Llm Chatbot | |
| emoji: π | |
| colorFrom: indigo | |
| colorTo: red | |
| sdk: gradio | |
| sdk_version: 5.4.0 | |
| app_file: app.py | |
| pinned: false | |
| license: apache-2.0 | |
| Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference | |
| # LLM-Based Chat Application | |
| This is a language model-based chat application that provides users with relevant, up-to-date, and easy-to-understand responses by leveraging web search and a powerful language model. The basic flow involves performing a web search using DuckDuckGo, followed by generating a response using the Qwen2.5-0.5b-Instruct language model. | |
| Live project: | |
| ## Features | |
| - **Web Search**: Performs real-time searches using DuckDuckGo to retrieve the latest relevant information. | |
| - **Language Model Integration**: Uses the `qwen2.5-0.5b-instruct` model for generating accurate and coherent responses. | |
| - **Simple Interface**: Built with Gradio to offer an easy-to-use web interface for users to interact with the model. | |
| - **Efficient and Scalable**: Optimized for performance using Intel-specific optimizations via `optimum-intel` and other acceleration libraries. | |
| ## Model Used | |
| - Language Model: qwen2.5-0.5b-instruct | |
| - A powerful and efficient model that generates human-like responses based on the query. | |
| ## Application Flow | |
| - **User Query**: The user inputs a question or request. | |
| - **Web Search**: The system performs a web search using DuckDuckGo to gather recent and relevant information. | |
| - **Response Generation**: The gathered data is used to generate a clear, relevant, and easy-to-understand response using the qwen2.5-0.5b-instruct language model. | |
| - **User Response**: The application returns the generated response to the user. | |
| ## Dependencies | |
| The following dependencies are required for running this application: | |
| - `openvino>=2024.2.0` | |
| - `openvino-tokenizers[transformers]` | |
| - `torch>=2.1` | |
| - `datasets` | |
| - `duckduckgo-search` | |
| - `langchain-community` | |
| - `accelerate` | |
| - `gradio>=4.19` | |
| - `onnx<=1.16.1` (For Windows platform `sys_platform=='win32'`) | |
| - `einops` | |
| - `transformers>=4.43.1` | |
| - `transformers_stream_generator` | |
| - `tiktoken` | |
| - `bitsandbytes` | |
| - `optimum-intel` (installed via `git+https://github.com/huggingface/optimum-intel.git`) | |
| - `nncf` (installed via `git+https://github.com/openvinotoolkit/nncf.git`) | |
| ## Installation | |
| To install the required dependencies, you can use the following command: | |
| ```bash | |
| pip install -r requirements.txt | |
| ``` | |
| ## License | |
| - This project is licensed under the Apache License 2.0. See the LICENSE file for details. | |
| ## Acknowledgements | |
| - **OpenVINO**: For efficient acceleration of the language model. | |
| - **DuckDuckGo**: For providing real-time web search capabilities. | |
| - **Hugging Face**: For providing powerful transformer models and tools. | |