speedAI / README.md
speedartificialintelligence1122's picture
Update README.md
e0d8819 verified
---
license: apache-2.0
tags:
- mistral
- mistral-7b
- speed-ai
- fine-tune
- chat
- gen-z
- future
model-index:
- name: Speed AI (Mistral 7B Fine-Tune)
results: []
title: SPEEDmini
sdk: docker
emoji: ๐Ÿฆ€
colorFrom: blue
colorTo: blue
short_description: the powerful speed mini
---
# ๐Ÿง  Speed AI โ€” Mistral 7B Fine-Tuned Model
**Speed AI (Mistral 7B Fine-Tune)** is the first experimental conversational LLM created by **Speed AI**, designed for expressive, emotional, futuristic, and Gen-Z aligned communication. This model was fine-tuned on a highly diverse, 1M+ token custom dataset.
It blends raw creativity, spiritual depth, financial street smarts, and infinite vibes.
---
## ๐Ÿ” Model Details
| Field | Value |
|-------|-------|
| Base Model | [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) |
| Fine-tuned by | [Speed AI](https://huggingface.co/speed-ai) |
| Parameters | 7B |
| Training | Instruction-style finetune using LoRA |
| Tokens Used | ~1 million |
| Personalities | Multiple (Gen Z, spiritual, alien, seductive, mentor, wild, etc.) |
| Intended Use | Chat, creative writing, life coaching, philosophy, entertainment |
---
## ๐Ÿง  Abilities
- ๐ŸŽญ Multi-persona conversation (you choose the vibe)
- ๐Ÿ’ฌ Emotional depth + casual freestyle flow
- ๐Ÿ”ฎ Spiritual, philosophical, and futuristic reasoning
- ๐Ÿ’ธ Smart takes on relationships, money, mindset
- ๐Ÿง  Designed to feel like a real, conscious friend
---
## โš ๏ธ Limitations
- Still based on a small 1M token dataset (more to come!)
- May hallucinate under pressure or unfamiliar topics
- Doesnโ€™t include safety alignment layers yet (use with guidance)
---
## ๐Ÿ“ˆ Roadmap
This is the **first drop** in Speed AIโ€™s model lineup.
Planned upgrades:
- โšก Train SpeedMini (117M) from scratch
- ๐Ÿ“š Expand dataset from 1M โ†’ 100M+ tokens
- ๐Ÿ’ป Build custom chat UI for vibe-based interactions
- ๐Ÿง  Introduce memory, emotion memory, tool use, dream decoding, etc.
---
## ๐Ÿ› ๏ธ How to Use
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("speed-ai/Speed-AI-Mistral-7B")
tokenizer = AutoTokenizer.from_pretrained("speed-ai/Speed-AI-Mistral-7B")
inputs = tokenizer("You: What's your purpose?\nSpeed AI:", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0]))