saishshinde15 commited on
Commit
5fa0c2d
·
verified ·
1 Parent(s): 242279a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -15,16 +15,16 @@ language:
15
  pipeline_tag: text-generation
16
  ---
17
 
18
- # TBH.AI Vortex Reasoning
19
 
20
- - **Developed by:** TBH.AI
21
  - **License:** apache-2.0
22
  - **Fine-tuned from:** [saishshinde15/TBH.AI_Base_Reasoning](https://huggingface.co/saishshinde15/TBH.AI_Base_Reasoning)
23
  - **Category:** Experimental, Research
24
 
25
  ## **Introduction**
26
 
27
- TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [TBH.AI Base Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning). While the Base Reasoning model utilized **Generalized Reinforced Policy Optimization (GRPO)** to enhance step-by-step logical thought processes similar to **DeepSeek-R1**, this model takes a different approach—**eliminating GRPO and instead relying on high-end Supervised Fine-Tuning (SFT) techniques**.
28
 
29
  The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
30
 
@@ -49,7 +49,7 @@ To compensate for the lack of reinforcement learning, we used an **extensive dat
49
  The model internally uses **special reasoning markers** (`<think>` and `<answer>`) to structure its responses, though these may not always be visible in the final output. This ensures a **consistent and methodical approach** to answering questions.
50
 
51
  ### **5️⃣ Part of the TethysAI Vortex Family**
52
- This model belongs to the **TBH.AI Vortex series**, a collection of fine-tuned models pushing the boundaries of **SFT-based reasoning without reinforcement learning**.
53
 
54
  ---
55
 
@@ -76,7 +76,7 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
76
  import torch
77
 
78
  # Load model & tokenizer
79
- model_name = "saishshinde15/TBH.AI_Vortex_Reasoning"
80
  tokenizer = AutoTokenizer.from_pretrained(model_name)
81
  model = AutoModelForCausalLM.from_pretrained(model_name).to("cuda")
82
 
 
15
  pipeline_tag: text-generation
16
  ---
17
 
18
+ # Clyrai Vortex Reasoning
19
 
20
+ - **Developed by:** clyrai
21
  - **License:** apache-2.0
22
  - **Fine-tuned from:** [saishshinde15/TBH.AI_Base_Reasoning](https://huggingface.co/saishshinde15/TBH.AI_Base_Reasoning)
23
  - **Category:** Experimental, Research
24
 
25
  ## **Introduction**
26
 
27
+ TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [Clyrai Base Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning). While the Base Reasoning model utilized **Generalized Reinforced Policy Optimization (GRPO)** to enhance step-by-step logical thought processes similar to **DeepSeek-R1**, this model takes a different approach—**eliminating GRPO and instead relying on high-end Supervised Fine-Tuning (SFT) techniques**.
28
 
29
  The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
30
 
 
49
  The model internally uses **special reasoning markers** (`<think>` and `<answer>`) to structure its responses, though these may not always be visible in the final output. This ensures a **consistent and methodical approach** to answering questions.
50
 
51
  ### **5️⃣ Part of the TethysAI Vortex Family**
52
+ This model belongs to the **Clyrai Vortex series**, a collection of fine-tuned models pushing the boundaries of **SFT-based reasoning without reinforcement learning**.
53
 
54
  ---
55
 
 
76
  import torch
77
 
78
  # Load model & tokenizer
79
+ model_name = "saishshinde15/Clyrai_Vortex_Reasoning"
80
  tokenizer = AutoTokenizer.from_pretrained(model_name)
81
  model = AutoModelForCausalLM.from_pretrained(model_name).to("cuda")
82