shaheerawan3's picture
Create READ.md
9d4b1b6 verified

A newer version of the Gradio SDK is available: 6.1.0

Upgrade

Mistral-7B-Instruct v0.3 Advanced Chatbot (Gradio Version)

This repository contains a comprehensive Gradio application that serves as an advanced chatbot powered by the Mistral-7B-Instruct v0.3 model from Hugging Face. It's optimized for deployment on Hugging Face Spaces.

Features

  • Interactive chat interface with Mistral-7B-Instruct v0.3
  • Multiple chat sessions management
  • Customizable system prompts
  • Adjustable generation parameters (temperature, max tokens, etc.)
  • File analysis for:
    • CSV files
    • Excel files
    • Text files
    • JSON files
  • Context-aware responses that can incorporate file data
  • JSON structure generation optimized for n8n workflows

Requirements

  • Python 3.8+
  • Required packages are listed in requirements.txt

Quick Start

  1. Clone this repository
  2. Install dependencies:
    pip install -r requirements.txt
    
  3. Run the application:
    python app.py
    

Deployment to Hugging Face Spaces

This application is designed to work well with Hugging Face Spaces:

  1. Create a new Space on Hugging Face (https://huggingface.co/spaces)
  2. Choose the "Gradio" framework
  3. Upload the files from this repository
  4. Make sure to select a GPU runtime for better performance

Usage Guide

  1. Start by clicking "Load Mistral-7B Model" in the interface
  2. Wait for the model to load completely (this may take a few minutes on first run)
  3. Type your message in the chat input field and press Send
  4. Create new chat sessions as needed
  5. Adjust system prompt and generation parameters for better results
  6. Upload files for analysis and incorporate their data in your prompts

Specialized for n8n JSON Generation

The default system prompt is optimized for generating well-structured JSON for n8n workflows. You can:

  1. Ask the model to create complex JSON structures
  2. Request specific n8n node configurations
  3. Generate sample data in the correct format
  4. Validate and fix existing JSON for n8n compatibility

Model Configuration

The application provides various configuration options:

  • System Prompt: Define how the AI should behave
  • Temperature: Control creativity (higher = more creative)
  • Max Tokens: Limit the length of responses
  • Top P: Nucleus sampling parameter
  • Repetition Penalty: Reduce repetition in responses

Notes

  • The first load of the model may take several minutes depending on your hardware
  • For Hugging Face Spaces, a GPU runtime is strongly recommended
  • The model requires approximately 14GB+ of VRAM for optimal performance

License

This project is released under the MIT License.