prithivMLmods's picture
Update README.md
d49edda verified
---
license: apache-2.0
language:
- en
- zh
base_model:
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
pipeline_tag: text-generation
library_name: transformers
tags:
- text-generation-inference
- math
- code
- reasoning
- R1
---
![vvvvvvvvvvv.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/cFNcCXQNciTqiIqEdtggJ.png)
# **Magellanic-Qwen-14B-R1**
> **Magellanic-Qwen-14B-R1** is based on the **DeepSeek-R1-Distill-Qwen-14B** modality architecture, enhanced specifically for **mathematical reasoning** and **coding reasoning**. This model advances the capabilities of 14B-parameter architectures, excelling in logic-based problem solving, programming tasks, and context-rich dialogue generation. It is fine-tuned with extended chain-of-thought reasoning and domain-specific datasets for improved comprehension, structured generation, and precision in technical tasks.
## **Key Improvements**
1. **Mathematical Reasoning Enhancements**
Optimized with datasets targeting arithmetic, algebra, calculus, and formal logic, improving step-by-step solution generation and explanation accuracy.
2. **Coding Reasoning Enhancements**
Fine-tuned on diverse programming languages and reasoning-based coding problems (e.g., LeetCode, Codeforces, and real-world engineering tasks), significantly improving code generation, debugging, and documentation.
3. **Enhanced General Knowledge**
Broad knowledge base across various domains enables accurate and coherent responses for diverse topics.
4. **Improved Instruction Following**
Better handling of complex, multi-step instructions with structured and logically coherent outputs.
5. **Versatile Adaptability**
Resilient across open-ended and structured prompts, adapting well to different interaction styles and subject areas.
6. **Long-Context Support**
Supports up to **128K tokens** of input context and can generate up to **8K tokens** of output—ideal for in-depth technical and academic outputs.
## **Quickstart with transformers**
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "prithivMLmods/Magellanic-Qwen-14B-R1"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Explain how quicksort works with an example in Python."
messages = [
{"role": "system", "content": "You are a helpful assistant skilled in coding and reasoning tasks."},
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=512
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
```
## **Intended Use**
1. **Mathematics and Logic Tasks**
Solve and explain math problems, logical puzzles, and formula-based reasoning tasks step-by-step.
2. **Programming and Development**
Assist in generating code, debugging, documenting functions, and solving algorithmic problems across multiple languages.
3. **General-Purpose Reasoning**
Handle a wide variety of questions with accurate, contextual responses based on general knowledge and logic.
4. **Educational Assistance**
Help students and educators with clear, structured explanations in STEM and non-STEM subjects.
5. **Conversational AI & Chatbots**
Power intelligent assistants that require contextual awareness and technically sound responses.
6. **Multilingual Applications**
Translate, summarize, and generate multilingual content for global users.
7. **Long-Form Content Generation**
Generate coherent long articles, research summaries, and reports, especially with structured technical content.
## **Limitations**
1. **High Resource Usage**
Requires high-memory GPUs/TPUs for efficient inference, especially when utilizing 128K context.
2. **Bias and Hallucination Risk**
May reflect biases from pretraining data and occasionally hallucinate plausible-sounding but incorrect facts.
3. **Variability in Creative Tasks**
Less consistent in producing high-quality creative writing or highly subjective content.
4. **Training Cutoff Constraints**
No access to real-world events beyond the last training snapshot.
5. **Error Propagation in Long Outputs**
Minor early mistakes can compound in very long outputs.
6. **Prompt Sensitivity**
Performance may vary depending on prompt clarity and structure.