Model Card for Gear-1-160M
Model Details
Model Description
Gear-1-160M is a small Transformer LLM with about 160 million parameters in GGUF format, designed to run fast on local machines with low memory (CPU/GPU). It’s good for simple chat and basic tasks, but it may be slow or make mistakes since this is my first attempt at making a neural network. I plan to improve it in the future, so please bear with it. A 300M-parameter version will be coming soon.
- Developed by: HeavensHack
- Model type: GPTNeoX (general.architecture = gptneox).
- Language(s) (NLP): English
- License: MIT
Uses
Direct Use
- Simple question answering
Out-of-Scope Use
- Complex reasoning tasks
- High-stakes decision making
- Content requiring very high accuracy
Bias, Risks, and Limitations
- May make mistakes or “hallucinate” answers
- First experimental version, so behavior can be inconsistent
- Not suitable for critical tasks
Recommendations
- Use for learning, demos, or hobby projects
- Do not rely on it for serious or sensitive tasks
How to Get Started
- Load the GGUF model in your preferred LLM inference framework (LMStudio)
- Designed for CPU/GPU local inference
Training Details
- Training Data: Custom dataset / small-scale experimental
- Training Procedure: First experimental training, not fully optimized
- Training Regime: Mixed precision (FP16/FP32)
Evaluation
- Tested on small chat prompts
- Can respond to simple queries
- Results may vary; not benchmarked for production use
Environmental Impact
- Small-scale training, low compute and low carbon footprint
Model Card Contact
But...
- At first, only the GGUF file will be available. The rest will follow later.