Gear-1-160m / README.md
HeavensHackDev's picture
Update README.md
5fd76d3 verified
---
language:
- en
pipeline_tag: text-generation
license: apache-2.0
---
# Model Card for Gear-1-160M
## Model Details
### Model Description
Gear-1-160M is a small Transformer LLM with about 160 million parameters in GGUF format, designed to run fast on local machines with low memory (CPU/GPU). It’s good for simple chat and basic tasks, but it may be slow or make mistakes since this is my first attempt at making a neural network. I plan to improve it in the future, so please bear with it. A 300M-parameter version will be coming soon.
- **Developed by:** HeavensHack
- **Model type:** GPTNeoX (general.architecture = gptneox).
- **Language(s) (NLP):** English
- **License:** MIT
## Uses
### Direct Use
- Simple question answering
### Out-of-Scope Use
- Complex reasoning tasks
- High-stakes decision making
- Content requiring very high accuracy
## Bias, Risks, and Limitations
- May make mistakes or “hallucinate” answers
- First experimental version, so behavior can be inconsistent
- Not suitable for critical tasks
### Recommendations
- Use for learning, demos, or hobby projects
- Do not rely on it for serious or sensitive tasks
## How to Get Started
- Load the GGUF model in your preferred LLM inference framework (LMStudio)
- Designed for CPU/GPU local inference
## Training Details
- **Training Data:** Custom dataset / small-scale experimental
- **Training Procedure:** First experimental training, not fully optimized
- **Training Regime:** Mixed precision (FP16/FP32)
## Evaluation
- Tested on small chat prompts
- Can respond to simple queries
- Results may vary; not benchmarked for production use
## Environmental Impact
- Small-scale training, low compute and low carbon footprint
## Model Card Contact
- **Author:** HeavensHackDev
# But...
- **At first, only the GGUF file will be available. The rest will follow later.**