utkarshshukla2912 commited on
Commit
79afe69
·
verified ·
1 Parent(s): 4e8e4db

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +96 -2
README.md CHANGED
@@ -10,6 +10,102 @@ language:
10
  - en
11
  ---
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  # Uploaded finetuned model
14
 
15
  - **Developed by:** RinggAI
@@ -17,5 +113,3 @@ language:
17
  - **Finetuned from model :** Qwen/Qwen3.5-0.8B
18
 
19
  This qwen3_5 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
-
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
10
  - en
11
  ---
12
 
13
+
14
+ The model was finetuned on ~128,000 curated transcripts across different domanins and language preferences
15
+ - Expanded Training: Now optimized for CX Support, Healthcare, Loan Collection, Insurance, Ecommerce, and Concierge services.
16
+ - Feature Improvement: Significantly enhanced relative date-time extraction for more precise data processing.
17
+ - Training Overview
18
+ - You can plug it into your calling or voice AI stack to automatically extract:
19
+ - Enum-based classifications (e.g., call outcome, intent, disposition)
20
+ - Conversation summaries
21
+ - Action items / follow-ups
22
+ - Relative DateTime Artifacts
23
+
24
+ It’s built to handle real-world Hindi, English, Indic noisy transcripts.
25
+
26
+ Finetuning Parameters:
27
+ ```
28
+ rank = 64 # kept small to know change inherent model intelligence but to make sure structured ectraction is followed
29
+ trainer = SFTTrainer(
30
+ model = model,
31
+ tokenizer = tokenizer,
32
+ train_dataset = train_dataset,
33
+ eval_dataset = test_dataset,
34
+ args = SFTConfig(
35
+ dataset_text_field = "prompt",
36
+ max_seq_length = max_seq_length,
37
+ per_device_train_batch_size = 5,
38
+ gradient_accumulation_steps = 5,
39
+
40
+ warmup_steps = 10,
41
+ num_train_epochs = 2,
42
+ learning_rate = 2e-4,
43
+ lr_scheduler_type = "linear",
44
+
45
+ optim = "adamw_8bit",
46
+ weight_decay = 0.01, # Unsloth default (was 0.001)
47
+ seed = SEED,
48
+
49
+ logging_steps = 50,
50
+ report_to = "wandb",
51
+
52
+ eval_strategy = "steps",
53
+ eval_steps = 5000,
54
+ save_strategy = "steps",
55
+ save_steps = 5000,
56
+ load_best_model_at_end = True,
57
+ metric_for_best_model = "eval_loss",
58
+
59
+ output_dir = "outputs_qwen35_0.8b",
60
+ dataset_num_proc = 8,
61
+ fp16= not torch.cuda.is_bf16_supported(),
62
+ bf16= torch.cuda.is_bf16_supported()
63
+ ),
64
+ )
65
+ ```
66
+ ![Training Overview](logs.jpg)
67
+
68
+ `PS: VERY FEW EVALs WERE TAKEN FOR THE 0.8b MODEL`
69
+
70
+
71
+ Provide the below schema for best output:
72
+ ```
73
+ response_schema = {
74
+ "type": "object",
75
+ "properties": {
76
+ "key_points": {
77
+ "type": "array",
78
+ "items": {"type": "string"},
79
+ "nullable": True,
80
+ },
81
+ "action_items": {
82
+ "type": "array",
83
+ "items": {"type": "string"},
84
+ "nullable": True,
85
+ },
86
+ "summary": {"type": "string"},
87
+ "classification": classification_schema,
88
+ "callback_requested": {
89
+ "type": "boolean",
90
+ "nullable": False,
91
+ "description": "If the user requested a callback or mentiones currently he is busy then value is true otherwise false",
92
+ },
93
+ "callback_requested_time": {
94
+ "type": "string",
95
+ "nullable": True,
96
+ "description": "ISO 8601 datetime string (YYYY-MM-DDTHH:MM:SS) in the call's timezone, if user requested a callback",
97
+ },
98
+ },
99
+ "required": ["summary", "classification"],
100
+ }
101
+ ```
102
+
103
+
104
+
105
+ [<img style="border-radius: 20px;" src="https://storage.googleapis.com/desivocal-prod/desi-vocal/logo.png" width="200"/>](https://ringg.ai)
106
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
107
+
108
+
109
  # Uploaded finetuned model
110
 
111
  - **Developed by:** RinggAI
 
113
  - **Finetuned from model :** Qwen/Qwen3.5-0.8B
114
 
115
  This qwen3_5 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.