|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: google/functiongemma-270m-it |
|
|
library_name: mlx |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- quantllm |
|
|
- mlx |
|
|
- mlx-lm |
|
|
- apple-silicon |
|
|
- transformers |
|
|
- q4_k_m |
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
# π functiongemma-270m-it-4bit-mlx |
|
|
|
|
|
**google/functiongemma-270m-it** converted to **MLX** format |
|
|
|
|
|
[](https://github.com/codewithdark-git/QuantLLM) |
|
|
[]() |
|
|
[]() |
|
|
|
|
|
<a href="https://github.com/codewithdark-git/QuantLLM">β Star QuantLLM on GitHub</a> |
|
|
|
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
## π About This Model |
|
|
|
|
|
This model is **[google/functiongemma-270m-it](https://huggingface.co/google/functiongemma-270m-it)** converted to **MLX** format optimized for Apple Silicon (M1/M2/M3/M4) Macs with native acceleration. |
|
|
|
|
|
| Property | Value | |
|
|
|----------|-------| |
|
|
| **Base Model** | [google/functiongemma-270m-it](https://huggingface.co/google/functiongemma-270m-it) | |
|
|
| **Format** | MLX | |
|
|
| **Quantization** | Q4_K_M | |
|
|
| **License** | apache-2.0 | |
|
|
| **Created With** | [QuantLLM](https://github.com/codewithdark-git/QuantLLM) | |
|
|
|
|
|
|
|
|
## π Quick Start |
|
|
|
|
|
### Generate Text with mlx-lm |
|
|
|
|
|
```python |
|
|
from mlx_lm import load, generate |
|
|
|
|
|
# Load the model |
|
|
model, tokenizer = load("QuantLLM/functiongemma-270m-it-4bit-mlx") |
|
|
|
|
|
# Simple generation |
|
|
prompt = "Explain quantum computing in simple terms" |
|
|
messages = [{"role": "user", "content": prompt}] |
|
|
prompt_formatted = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
add_generation_prompt=True |
|
|
) |
|
|
|
|
|
# Generate response |
|
|
text = generate(model, tokenizer, prompt=prompt_formatted, verbose=True) |
|
|
print(text) |
|
|
``` |
|
|
|
|
|
### Streaming Generation |
|
|
|
|
|
```python |
|
|
from mlx_lm import load, stream_generate |
|
|
|
|
|
model, tokenizer = load("QuantLLM/functiongemma-270m-it-4bit-mlx") |
|
|
|
|
|
prompt = "Write a haiku about coding" |
|
|
messages = [{"role": "user", "content": prompt}] |
|
|
prompt_formatted = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
add_generation_prompt=True |
|
|
) |
|
|
|
|
|
# Stream tokens as they're generated |
|
|
for token in stream_generate(model, tokenizer, prompt=prompt_formatted, max_tokens=200): |
|
|
print(token, end="", flush=True) |
|
|
``` |
|
|
|
|
|
### Command Line Interface |
|
|
|
|
|
```bash |
|
|
# Install mlx-lm |
|
|
pip install mlx-lm |
|
|
|
|
|
# Generate text |
|
|
python -m mlx_lm.generate --model QuantLLM/functiongemma-270m-it-4bit-mlx --prompt "Hello!" |
|
|
|
|
|
# Interactive chat |
|
|
python -m mlx_lm.chat --model QuantLLM/functiongemma-270m-it-4bit-mlx |
|
|
``` |
|
|
|
|
|
### System Requirements |
|
|
|
|
|
| Requirement | Minimum | |
|
|
|-------------|---------| |
|
|
| **Chip** | Apple Silicon (M1/M2/M3/M4) | |
|
|
| **macOS** | 13.0 (Ventura) or later | |
|
|
| **Python** | 3.10+ | |
|
|
| **RAM** | 8GB+ (16GB recommended) | |
|
|
|
|
|
```bash |
|
|
# Install dependencies |
|
|
pip install mlx-lm |
|
|
``` |
|
|
|
|
|
|
|
|
## π Model Details |
|
|
|
|
|
| Property | Value | |
|
|
|----------|-------| |
|
|
| **Original Model** | [google/functiongemma-270m-it](https://huggingface.co/google/functiongemma-270m-it) | |
|
|
| **Format** | MLX | |
|
|
| **Quantization** | Q4_K_M | |
|
|
| **License** | `apache-2.0` | |
|
|
| **Export Date** | 2025-12-21 | |
|
|
| **Exported By** | [QuantLLM v2.0](https://github.com/codewithdark-git/QuantLLM) | |
|
|
|
|
|
|
|
|
|
|
|
--- |
|
|
|
|
|
## π Created with QuantLLM |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
[](https://github.com/codewithdark-git/QuantLLM) |
|
|
|
|
|
**Convert any model to GGUF, ONNX, or MLX in one line!** |
|
|
|
|
|
```python |
|
|
from quantllm import turbo |
|
|
|
|
|
# Load any HuggingFace model |
|
|
model = turbo("google/functiongemma-270m-it") |
|
|
|
|
|
# Export to any format |
|
|
model.export("mlx", quantization="Q4_K_M") |
|
|
|
|
|
# Push to HuggingFace |
|
|
model.push("your-repo", format="mlx") |
|
|
``` |
|
|
|
|
|
<a href="https://github.com/codewithdark-git/QuantLLM"> |
|
|
<img src="https://img.shields.io/github/stars/codewithdark-git/QuantLLM?style=social" alt="GitHub Stars"> |
|
|
</a> |
|
|
|
|
|
**[π Documentation](https://github.com/codewithdark-git/QuantLLM#readme)** Β· |
|
|
**[π Report Issue](https://github.com/codewithdark-git/QuantLLM/issues)** Β· |
|
|
**[π‘ Request Feature](https://github.com/codewithdark-git/QuantLLM/issues)** |
|
|
|
|
|
</div> |
|
|
|