geometry-model-hf / README.md
aryan27's picture
Update README.md
87e6dd2 verified
---
license: mit
datasets:
- aryan27/geometry-cot
language:
- en
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
library_name: transformers
---
# **MiniLLM Geometry Engine**
MiniLLM Geometry Engine is a fine-tuned version of the TinyLlama 1.1B
language model, optimized for generating geometry-related functions for
a geometry engine. Fine-tuned on a custom Geometry Chain-of-Thought
(CoT) dataset, this model excels at producing accurate and efficient
mathematical functions for geometric computations, stored in the Hugging
Face .safetensors format for secure and efficient model loading.
## **Model Overview**
- **Base Model**: TinyLlama 1.1B
- **Fine-Tuning Dataset**: Custom Geometry CoT dataset consisting of coordinate geometry problems and geometrical construction instruction
- **Model Format**: .safetensors for secure and efficient weight storage
- **Intended Use**: Generating geometry functions for educational tools.
- **License**: MIT
- **Contact**: Contact via Hugging Face profile
## **Important Note:**
- Use the system_prompt.txt for the System Prompt, this provides the accurate results.
- You can extract and edit the system prompt. I hope to add new functions later.
## **Capabilities**
MiniLLM Geometry Engine interprets geometry-related queries and outputs
functional Python code or mathematical expressions. For example, given a
prompt like \"Generate a function to calculate the area of a triangle,\"
it produces executable code or formulas with clear reasoning, leveraging
its CoT fine-tuning for logical accuracy. Supported tasks include area,
volume, distance, and intersection calculations.
## **Technical Details**
- **Architecture**: Transformer-based, inherited from TinyLlama 1.1B
- **Fine-Tuning Details**: Trained on a dataset of geometry problems with step-by-step solutions
- **Output Format**: Python code or pseudocode compatible with geometry engine APIs
- **Performance**: Improved accuracy on geometry tasks compared to the base TinyLlama model, with low latency
- **Model Storage**: Uses .safetensors for secure and efficient loading with Hugging Face\'s safetensors library
## **Usage**
To load and use the model with Hugging Face\'s Transformers library:
from transformers import AutoModelForCausalLM, AutoTokenizer
from safetensors.torch import load_file
\# Load tokenizer and model
```python
tokenizer = AutoTokenizer.from_pretrained("aryan27/geometry-model-hf")
model = AutoModelForCausalLM.from_pretrained("aryan27/geometry-model-hf")
```
\# Example prompt
```python
prompt = "Draw a point A at 3,4"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(*inputs, max_length=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
Install required libraries:
```shell
pip install safetensors transformers
```
## **Training Details**
- **Dataset**: Custom Geometry CoT dataset with 10,000 geometry problems and solutions
- **Training Procedure**: Fine-tuned for 3 epochs on a single NVIDIA A100 GPU
- **Hyperparameters**: Learning rate: 2e-5, batch size: 16
- **Hardware**: NVIDIA A100 GPU
## **Limitations**
- Supports very few functions since this was an experimental model, maybe I will add other functions later.
- Limited to geometry-related tasks and may not generalize to other mathematical domains.
- Extensive use of strict System Prompting. I aim to eliminate that also.
## **How to Contribute**
Submit issues or pull requests via the Hugging Face repository.
Contributions to the dataset or model improvements are welcome.
## **Acknowledgments**
Thanks to the TinyLlama team for the base model and Hugging Face for the
safetensors and transformers libraries.
MiniLLM Geometry Engine is a lightweight, efficient solution for
developers and researchers needing reliable geometry function
generation, with the security and performance benefits of the
.safetensors format.