| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | base_model: prithivMLmods/QwQ-R1-Distill-7B-CoT |
| | pipeline_tag: text-generation |
| | library_name: transformers |
| | tags: |
| | - text-generation-inference |
| | - mlx |
| | - mlx-my-repo |
| | model-index: |
| | - name: QwQ-R1-Distill-7B-CoT |
| | results: |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: IFEval (0-Shot) |
| | type: wis-k/instruction-following-eval |
| | split: train |
| | args: |
| | num_few_shot: 0 |
| | metrics: |
| | - type: inst_level_strict_acc and prompt_level_strict_acc |
| | value: 35.0 |
| | name: averaged accuracy |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: BBH (3-Shot) |
| | type: SaylorTwift/bbh |
| | split: test |
| | args: |
| | num_few_shot: 3 |
| | metrics: |
| | - type: acc_norm |
| | value: 20.95 |
| | name: normalized accuracy |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: MATH Lvl 5 (4-Shot) |
| | type: lighteval/MATH-Hard |
| | split: test |
| | args: |
| | num_few_shot: 4 |
| | metrics: |
| | - type: exact_match |
| | value: 27.19 |
| | name: exact match |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: GPQA (0-shot) |
| | type: Idavidrein/gpqa |
| | split: train |
| | args: |
| | num_few_shot: 0 |
| | metrics: |
| | - type: acc_norm |
| | value: 5.82 |
| | name: acc_norm |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: MuSR (0-shot) |
| | type: TAUR-Lab/MuSR |
| | args: |
| | num_few_shot: 0 |
| | metrics: |
| | - type: acc_norm |
| | value: 4.5 |
| | name: acc_norm |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | - task: |
| | type: text-generation |
| | name: Text Generation |
| | dataset: |
| | name: MMLU-PRO (5-shot) |
| | type: TIGER-Lab/MMLU-Pro |
| | config: main |
| | split: test |
| | args: |
| | num_few_shot: 5 |
| | metrics: |
| | - type: acc |
| | value: 20.05 |
| | name: accuracy |
| | source: |
| | url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=prithivMLmods%2FQwQ-R1-Distill-7B-CoT |
| | name: Open LLM Leaderboard |
| | --- |
| | |
| | # nhe-ai/QwQ-R1-Distill-7B-CoT-mlx-2Bit |
| |
|
| | The Model [nhe-ai/QwQ-R1-Distill-7B-CoT-mlx-2Bit](https://huggingface.co/nhe-ai/QwQ-R1-Distill-7B-CoT-mlx-2Bit) was converted to MLX format from [prithivMLmods/QwQ-R1-Distill-7B-CoT](https://huggingface.co/prithivMLmods/QwQ-R1-Distill-7B-CoT) using mlx-lm version **0.22.1**. |
| |
|
| | ## Use with mlx |
| |
|
| | ```bash |
| | pip install mlx-lm |
| | ``` |
| |
|
| | ```python |
| | from mlx_lm import load, generate |
| | |
| | model, tokenizer = load("nhe-ai/QwQ-R1-Distill-7B-CoT-mlx-2Bit") |
| | |
| | prompt="hello" |
| | |
| | if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: |
| | messages = [{"role": "user", "content": prompt}] |
| | prompt = tokenizer.apply_chat_template( |
| | messages, tokenize=False, add_generation_prompt=True |
| | ) |
| | |
| | response = generate(model, tokenizer, prompt=prompt, verbose=True) |
| | ``` |
| |
|