YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
GodeusAI - Conversations with the infinite, locally hosted.
GodeusAI is a spiritual assistant chatbot powered by a fine-tuned Mistral-7B language model with LoRA adapters. It provides answers to spiritual and philosophical questions, and can be run locally for private, offline use.
Features
- Chatbot interface for spiritual Q&A
- Fine-tuned on custom spiritual/philosophical data
- LoRA adapter for efficient model adaptation
- Utilities for context-based question answering
- Example notebook for data preparation and training
Quickstart
1. Install Dependencies
pip install -r requirements.txt
2. Run the Chatbot
python inference.py
You’ll see a prompt:💉 Welcome to GodeusAI — your spiritual assistant. Type 'exit' to quit.
Project Structure
| File/Folder | Purpose |
|---|---|
inference.py |
Main script to chat with the model |
rag_utils.py |
Utilities for loading model and context-based QA |
few_shots_qa.jsonl |
Example Q&A pairs for few-shot prompting |
adapter_model.safetensors |
LoRA adapter weights (required for inference) |
adapter_config.json |
Configuration for the LoRA adapter |
GodeusAI.ipynb |
Notebook for data prep, training, and experimentation |
requirements.txt |
Python dependencies |
Data & Training
- few_shots_qa.jsonl: Example format for few-shot Q&A pairs.
- GodeusAI.ipynb:
- Prepares data from PDFs
- Chunks and formats into JSONL
- Trains LoRA adapters on Mistral-7B
Model
- Base model:
mistralai/Mistral-7B-v0.1 - Adapter: LoRA (config in
adapter_config.json, weights inadapter_model.safetensors)
License
This project is for research and educational purposes.
Refer to the base model and dataset licenses for usage restrictions.
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support