Files changed (1) hide show
  1. README.md +71 -59
README.md CHANGED
@@ -1,60 +1,72 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- base_model:
6
- - Qwen/Qwen2.5-1.5B-Instruct
7
- pipeline_tag: text-generation
8
- library_name: transformers
9
- tags:
10
- - trl
11
- - Reinforcement learning
12
- ---
13
- # **Bellatrix-Tiny-1.5B-R1**
14
-
15
- Bellatrix is based on a reasoning-based model designed for the DeepSeek-R1 synthetic dataset entries. The pipeline's instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. These models outperform many of the available open-source options. Bellatrix is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions utilize supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF).
16
-
17
- # **Use with transformers**
18
-
19
- Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.
20
-
21
- Make sure to update your transformers installation via `pip install --upgrade transformers`.
22
-
23
- ```python
24
- import torch
25
- from transformers import pipeline
26
-
27
- model_id = "prithivMLmods/Bellatrix-Tiny-1.5B-R1"
28
- pipe = pipeline(
29
- "text-generation",
30
- model=model_id,
31
- torch_dtype=torch.bfloat16,
32
- device_map="auto",
33
- )
34
- messages = [
35
- {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
36
- {"role": "user", "content": "Who are you?"},
37
- ]
38
- outputs = pipe(
39
- messages,
40
- max_new_tokens=256,
41
- )
42
- print(outputs[0]["generated_text"][-1])
43
- ```
44
-
45
- Note: You can also find detailed recipes on how to use the model locally, with `torch.compile()`, assisted generations, quantized and more at [`huggingface-llama-recipes`](https://github.com/huggingface/huggingface-llama-recipes)
46
-
47
- # **Intended Use**
48
- Bellatrix is designed for applications that require advanced reasoning and multilingual dialogue capabilities. It is particularly suitable for:
49
- - **Agentic Retrieval**: Enabling intelligent retrieval of relevant information in a dialogue or query-response system.
50
- - **Summarization Tasks**: Condensing large bodies of text into concise summaries for easier comprehension.
51
- - **Multilingual Use Cases**: Supporting conversations in multiple languages with high accuracy and coherence.
52
- - **Instruction-Based Applications**: Following complex, context-aware instructions to generate precise outputs in a variety of scenarios.
53
-
54
- # **Limitations**
55
- Despite its capabilities, Bellatrix has some limitations:
56
- 1. **Domain Specificity**: While it performs well on general tasks, its performance may degrade with highly specialized or niche datasets.
57
- 2. **Dependence on Training Data**: It is only as good as the quality and diversity of its training data, which may lead to biases or inaccuracies.
58
- 3. **Computational Resources**: The model’s optimized transformer architecture can be resource-intensive, requiring significant computational power for fine-tuning and inference.
59
- 4. **Language Coverage**: While multilingual, some languages or dialects may have limited support or lower performance compared to widely used ones.
 
 
 
 
 
 
 
 
 
 
 
 
60
  5. **Real-World Contexts**: It may struggle with understanding nuanced or ambiguous real-world scenarios not covered during training.
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-1.5B-Instruct
19
+ pipeline_tag: text-generation
20
+ library_name: transformers
21
+ tags:
22
+ - trl
23
+ - Reinforcement learning
24
+ ---
25
+ # **Bellatrix-Tiny-1.5B-R1**
26
+
27
+ Bellatrix is based on a reasoning-based model designed for the DeepSeek-R1 synthetic dataset entries. The pipeline's instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. These models outperform many of the available open-source options. Bellatrix is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions utilize supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF).
28
+
29
+ # **Use with transformers**
30
+
31
+ Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.
32
+
33
+ Make sure to update your transformers installation via `pip install --upgrade transformers`.
34
+
35
+ ```python
36
+ import torch
37
+ from transformers import pipeline
38
+
39
+ model_id = "prithivMLmods/Bellatrix-Tiny-1.5B-R1"
40
+ pipe = pipeline(
41
+ "text-generation",
42
+ model=model_id,
43
+ torch_dtype=torch.bfloat16,
44
+ device_map="auto",
45
+ )
46
+ messages = [
47
+ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
48
+ {"role": "user", "content": "Who are you?"},
49
+ ]
50
+ outputs = pipe(
51
+ messages,
52
+ max_new_tokens=256,
53
+ )
54
+ print(outputs[0]["generated_text"][-1])
55
+ ```
56
+
57
+ Note: You can also find detailed recipes on how to use the model locally, with `torch.compile()`, assisted generations, quantized and more at [`huggingface-llama-recipes`](https://github.com/huggingface/huggingface-llama-recipes)
58
+
59
+ # **Intended Use**
60
+ Bellatrix is designed for applications that require advanced reasoning and multilingual dialogue capabilities. It is particularly suitable for:
61
+ - **Agentic Retrieval**: Enabling intelligent retrieval of relevant information in a dialogue or query-response system.
62
+ - **Summarization Tasks**: Condensing large bodies of text into concise summaries for easier comprehension.
63
+ - **Multilingual Use Cases**: Supporting conversations in multiple languages with high accuracy and coherence.
64
+ - **Instruction-Based Applications**: Following complex, context-aware instructions to generate precise outputs in a variety of scenarios.
65
+
66
+ # **Limitations**
67
+ Despite its capabilities, Bellatrix has some limitations:
68
+ 1. **Domain Specificity**: While it performs well on general tasks, its performance may degrade with highly specialized or niche datasets.
69
+ 2. **Dependence on Training Data**: It is only as good as the quality and diversity of its training data, which may lead to biases or inaccuracies.
70
+ 3. **Computational Resources**: The model’s optimized transformer architecture can be resource-intensive, requiring significant computational power for fine-tuning and inference.
71
+ 4. **Language Coverage**: While multilingual, some languages or dialects may have limited support or lower performance compared to widely used ones.
72
  5. **Real-World Contexts**: It may struggle with understanding nuanced or ambiguous real-world scenarios not covered during training.