import gradio as gr from transformers import AutoTokenizer, AutoModelForCausalLM import torch from gpt4all import GPT4All import os # Specify the local path to the downloaded model file model_path = "https://huggingface.co/spaces/DR-Rakshitha/wizardlm_api/blob/main/wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin" # Check if the model file exists locally if not os.path.exists(model_path): raise FileNotFoundError(f"Model file not found at {model_path}. Please download it manually.") # Initialize the GPT4All model model = GPT4All(model_path) def generate_text(input_text): output = model.generate(input_text) return output text_generation_interface = gr.Interface( fn=generate_text, inputs=[ gr.inputs.Textbox(label="Input Text"), ], outputs=gr.inputs.Textbox(label="Generated Text"), title="Falcon-7B Instruct", ).launch()