Gear-1-160m / README.md
HeavensHackDev's picture
Update README.md
5fd76d3 verified
metadata
language:
  - en
pipeline_tag: text-generation
license: apache-2.0

Model Card for Gear-1-160M

Model Details

Model Description

Gear-1-160M is a small Transformer LLM with about 160 million parameters in GGUF format, designed to run fast on local machines with low memory (CPU/GPU). It’s good for simple chat and basic tasks, but it may be slow or make mistakes since this is my first attempt at making a neural network. I plan to improve it in the future, so please bear with it. A 300M-parameter version will be coming soon.

  • Developed by: HeavensHack
  • Model type: GPTNeoX (general.architecture = gptneox).
  • Language(s) (NLP): English
  • License: MIT

Uses

Direct Use

  • Simple question answering

Out-of-Scope Use

  • Complex reasoning tasks
  • High-stakes decision making
  • Content requiring very high accuracy

Bias, Risks, and Limitations

  • May make mistakes or “hallucinate” answers
  • First experimental version, so behavior can be inconsistent
  • Not suitable for critical tasks

Recommendations

  • Use for learning, demos, or hobby projects
  • Do not rely on it for serious or sensitive tasks

How to Get Started

  • Load the GGUF model in your preferred LLM inference framework (LMStudio)
  • Designed for CPU/GPU local inference

Training Details

  • Training Data: Custom dataset / small-scale experimental
  • Training Procedure: First experimental training, not fully optimized
  • Training Regime: Mixed precision (FP16/FP32)

Evaluation

  • Tested on small chat prompts
  • Can respond to simple queries
  • Results may vary; not benchmarked for production use

Environmental Impact

  • Small-scale training, low compute and low carbon footprint

Model Card Contact

  • Author: HeavensHackDev

But...

  • At first, only the GGUF file will be available. The rest will follow later.