| | --- |
| | library_name: transformers |
| | license: apache-2.0 |
| | license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE |
| | pipeline_tag: text-generation |
| | base_model: |
| | - Qwen/Qwen3-0.6B |
| | --- |
| | |
| | # Qwen3-0.6B (Modified) |
| |
|
| | A fork of [Qwen/Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B), modified for use as a training target model with [PrimeIntellect-ai/verifiers](https://github.com/PrimeIntellect-ai/verifiers). |
| |
|
| | ## Changes |
| |
|
| | - Extracted `chat_template` from `tokenizer_config.json` into a separate `chat_template.jinja` file (latest transformers format) |
| | - Reversed thinking tag logic to enable thinking mode by default (`enable_thinking=True`) |
| |
|
| | ## Original Model |
| |
|
| | For model architecture, performance, and usage details, refer to [Qwen/Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B). |
| |
|