Safetensors
qwen2
FINGU-AI commited on
Commit
3870ca4
·
verified ·
1 Parent(s): 37c8b8e

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +84 -0
README.md ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - FINGU-AI/llm_evaluation_full_dataset_unique_both_directions_7_to_10_scores
5
+ language:
6
+ - en
7
+ - ko
8
+ - ja
9
+ - id
10
+ - zh
11
+ - my
12
+ - bn
13
+ - km
14
+ - mn
15
+ - ne
16
+ - ru
17
+ - uz
18
+ - tl
19
+ - pt
20
+ - si
21
+ - vi
22
+ ---
23
+ # QWEN2.5-32B-2600s-FP8: Advanced Multilingual Translation Model
24
+
25
+ ## Overview
26
+ **FINGU-AI/QWEN2.5-32B-2600s-FP8** is a fine-tuned version of Qwen 2.5 32B, specifically optimized for multilingual translation across **16 different languages**. This model has been extensively fine-tuned to enhance its translation capabilities, making it competitive with high-tier models like 72B in terms of translation accuracy and fluency.
27
+
28
+ ## Fine-Tuning Process
29
+ ### Data Collection
30
+ To improve the model's understanding and translation capabilities, we curated and synthesized a large dataset consisting of:
31
+ - High-quality multilingual conversational datasets.
32
+ - Real-world dialogues spanning general, business, and technical domains.
33
+ - Translated datasets covering diverse linguistic structures and idiomatic expressions.
34
+
35
+ ### Multilingual Enhancement
36
+ To advance its translation capabilities, we leveraged:
37
+ - **Translation Expansion**: The collected dataset was translated into **16 different languages** to ensure robust multilingual performance.
38
+ - **Benchmarking Against High-Tier Models**: We utilized state-of-the-art translation models, including **Gemini** and other top-ranking translation models with high BLEU and COMET scores, to refine our translation quality.
39
+ - **Reinforcement Learning with Human Feedback (RLHF)**: Translation outputs were evaluated and iteratively improved based on feedback from native speakers and linguistic experts.
40
+
41
+ ### Training and Optimization
42
+ - **Base Model**: Qwen 2.5 32B FP8
43
+ - **Fine-Tuning Framework**: LoRA + QLoRA for efficient training
44
+ - **Batch Size**: Optimized for multi-GPU environments
45
+ - **Precision**: FP8 for efficient computation without sacrificing performance
46
+ - **Training Iterations**: Over 2600 steps on **multi-H100 GPUs**
47
+
48
+ ## Key Improvements
49
+ - **Enhanced Multilingual Translation**: The model now achieves translation fluency comparable to 72B models across multiple language pairs.
50
+ - **Diverse Conversational Understanding**: Improved ability to process and generate accurate translations for various contexts, including business, casual, and formal speech.
51
+ - **Optimized for Low-Latency Inference**: Fine-tuned with efficiency in mind, making it suitable for real-time translation applications.
52
+
53
+ ## Performance Evaluation
54
+ The model was evaluated using:
55
+ - **BLEU, COMET, and chrF scores**: To measure translation quality across multiple languages.
56
+ - **Human Evaluation**: Involving bilingual speakers and linguistic professionals to validate accuracy and fluency.
57
+ - **Comparisons with SOTA Models**: Benchmarked against high-performance models like GPT-4, Gemini, and LLaMA-3 to ensure top-tier translation quality.
58
+
59
+ ## Usage
60
+ This model is suitable for:
61
+ - High-quality machine translation across multiple languages
62
+ - Conversational AI with multilingual capabilities
63
+ - Cross-lingual content generation and customer support
64
+ - NLP applications requiring robust and accurate translation
65
+
66
+ ## Limitations
67
+ - While translation quality is highly competitive, niche dialects or highly technical documents may require additional fine-tuning.
68
+ - Performance may vary slightly depending on the deployment environment and inference settings.
69
+
70
+ ## Citation
71
+ If you use this model, please cite:
72
+ ```
73
+ @misc{FINGU-AI-QWEN2.5-32B-2600s-FP8,
74
+ author = {FINGU-AI},
75
+ title = {FINGU-AI/QWEN2.5-32B-2600s-FP8: Advanced Multilingual Translation Model},
76
+ year = {2025},
77
+ publisher = {Hugging Face},
78
+ url = {https://huggingface.co/FINGU-AI/QWEN2.5-32B-2600s-FP8}
79
+ }
80
+ ```
81
+
82
+ ---
83
+ ### License
84
+ This model follows the licensing terms of the original Qwen 2.5 32B model. Ensure compliance with regional translation regulations before deploying in production environments.