--- language: - en pipeline_tag: text-generation license: apache-2.0 --- # Model Card for Gear-1-160M ## Model Details ### Model Description Gear-1-160M is a small Transformer LLM with about 160 million parameters in GGUF format, designed to run fast on local machines with low memory (CPU/GPU). It’s good for simple chat and basic tasks, but it may be slow or make mistakes since this is my first attempt at making a neural network. I plan to improve it in the future, so please bear with it. A 300M-parameter version will be coming soon. - **Developed by:** HeavensHack - **Model type:** GPTNeoX (general.architecture = gptneox). - **Language(s) (NLP):** English - **License:** MIT ## Uses ### Direct Use - Simple question answering ### Out-of-Scope Use - Complex reasoning tasks - High-stakes decision making - Content requiring very high accuracy ## Bias, Risks, and Limitations - May make mistakes or “hallucinate” answers - First experimental version, so behavior can be inconsistent - Not suitable for critical tasks ### Recommendations - Use for learning, demos, or hobby projects - Do not rely on it for serious or sensitive tasks ## How to Get Started - Load the GGUF model in your preferred LLM inference framework (LMStudio) - Designed for CPU/GPU local inference ## Training Details - **Training Data:** Custom dataset / small-scale experimental - **Training Procedure:** First experimental training, not fully optimized - **Training Regime:** Mixed precision (FP16/FP32) ## Evaluation - Tested on small chat prompts - Can respond to simple queries - Results may vary; not benchmarked for production use ## Environmental Impact - Small-scale training, low compute and low carbon footprint ## Model Card Contact - **Author:** HeavensHackDev # But... - **At first, only the GGUF file will be available. The rest will follow later.**