compact-ai-model / model_card.md
likhonsheikh's picture
Upload folder using huggingface_hub
b9b1e87 verified
---
language: en
license: apache-2.0
tags:
- compact-ai
- interleaved-thinking
- transformer
- pytorch
- reasoning
datasets:
- custom
---
# Compact AI Model with Interleaved Thinking
A compact AI model that implements interleaved thinking for enhanced reasoning capabilities. This model combines efficient transformer architecture with parallel reasoning paths to achieve better performance on complex tasks.
## Model Details
### Model Description
This is a compact AI model designed for efficient inference while maintaining strong reasoning capabilities through interleaved thinking. The model uses multiple parallel reasoning paths that work together to solve complex problems.
### Model Architecture
- **Base Architecture**: Transformer with efficient attention mechanisms
- **Key Features**:
- Interleaved thinking with parallel reasoning paths
- Hierarchical reasoning with different abstraction levels
- Adaptive memory compression
- Early stopping based on confidence thresholds
- RoPE positional embeddings
- Flash attention support
### Model Sizes
- **Tiny**: ~50M parameters (256 dim, 8 layers, 8 heads)
- **Small**: ~100M parameters (512 dim, 12 layers, 8 heads)
- **Medium**: ~200M parameters (768 dim, 16 layers, 12 heads)
## Usage
### Installation
```bash
pip install torch transformers
```
### Loading the Model
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("likhonsheikh/compact-ai-model")
tokenizer = AutoTokenizer.from_pretrained("likhonsheikh/compact-ai-model")
```
### Inference
```python
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
```
### API Usage
The model also supports a FastAPI-based API server:
```bash
uvicorn compact_ai_model.api.main:app --host 0.0.0.0 --port 8000
```
## Training
### Requirements
- Python 3.8+
- PyTorch 2.0+
- CUDA-compatible GPU (recommended)
### Training Script
```bash
python compact_ai_model/training/train.py
```
## Performance
### Benchmarks
- **MMLU**: Coming soon
- **ARC**: Coming soon
- **HellaSwag**: Coming soon
### Efficiency
- Memory-efficient attention mechanisms
- Adaptive compression for long contexts
- Early stopping to reduce computation
## Limitations
- Currently uses a simple tokenizer for demonstration
- Model is not yet fine-tuned on large datasets
- API is still in development
## Citation
```bibtex
@misc{compact-ai-model,
title={Compact AI Model with Interleaved Thinking},
author={Likhon Sheikh},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/likhonsheikh/compact-ai-model}
}
```
## License
This model is released under the Apache 2.0 license.