sd-inf commited on
Commit
08cf508
·
verified ·
1 Parent(s): 153a6fa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -29
README.md CHANGED
@@ -1,46 +1,75 @@
 
 
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: Qwen/Qwen3-4b-thinking-2507
4
  tags:
5
- - peft
6
- - lora
7
- - merged
8
- - reasoning
 
9
  library_name: transformers
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
- # Astral-400
 
 
 
 
 
13
 
14
- This model is a merge of [Qwen/Qwen3-4b-thinking-2507](Qwen/Qwen3-4b-thinking-2507) with a LoRA adapter trained on reasoning data.
15
 
16
- ## Model Details
17
 
18
- - **Base Model**: Qwen/Qwen3-4b-thinking-2507
19
- - **Training**: LoRA fine-tuning on multi-domain reasoning tasks
20
- - **Merged**: Yes, this is the merged model (no adapter needed)
21
 
22
- ## Usage
 
 
 
 
 
 
23
 
24
- ```python
25
- from transformers import AutoModelForCausalLM, AutoTokenizer
26
 
27
- model = AutoModelForCausalLM.from_pretrained(
28
- "LucidityAI/Astral-400",
29
- torch_dtype="auto",
30
- device_map="auto"
31
- )
32
- tokenizer = AutoTokenizer.from_pretrained("LucidityAI/Astral-400")
33
 
34
- # Use the model for inference
35
- inputs = tokenizer("Your prompt here", return_tensors="pt")
36
- outputs = model.generate(**inputs, max_new_tokens=100)
37
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
38
- print(response)
 
 
 
 
 
 
 
39
  ```
40
 
41
- ## Training Data
42
 
43
- Trained on the [Astral Post-Training Dataset](https://huggingface.co/datasets/LucidityAI/Astral-Post-Training-Dataset) containing:
44
- - Mathematical reasoning problems
45
- - Code generation tasks
46
- - Scientific reasoning questions
 
 
 
 
1
+ Certainly! Here's a more professional, polished, and clearly structured version of your model card—refined for tone, clarity, and technical precision while preserving all original information:
2
+
3
  ---
4
+
5
+ ```yaml
6
  license: apache-2.0
7
  base_model: Qwen/Qwen3-4b-thinking-2507
8
  tags:
9
+ - reasoning
10
+ - code
11
+ - inference
12
+ - chat
13
+ - multimodal (if applicable)
14
  library_name: transformers
15
+ datasets:
16
+ - nvidia/AceReason-1.1-SFT
17
+ ```
18
+
19
+ # Astral-4B-Preview
20
+
21
+
22
+ Astral-4B is a specialized reasoning-focused language model developed as part of the Astral series, designed to deliver high-fidelity, step-by-step reasoning with configurable depth. Built upon the Qwen3-4b-thinking-2507 foundation, this variant has been fine-tuned on the `nvidia/AceReason-1.1-SFT` dataset to enhance logical coherence, problem-solving capability, and structured thinking.
23
+
24
+ This model is currently in **preview** and intended for research, evaluation, and development use. Feedback is encouraged to guide future iterations.
25
+
26
  ---
27
 
28
+ ## Key Features
29
+
30
+ - **Configurable Reasoning Depth**: Supports explicit control over reasoning granularity via system-level prompts.
31
+ - **ChatML Template Compliance**: Fully compatible with the ChatML format for seamless integration into existing inference pipelines.
32
+ - **Structured Output**: Generates detailed reasoning traces (`<think>` blocks) when enabled, enabling auditability and interpretability.
33
+ - **Optimized for Complex Tasks**: Strong performance on mathematical reasoning, code generation, and multi-step logic problems.
34
 
35
+ ---
36
 
37
+ ## Usage Instructions
38
 
39
+ To invoke the model correctly, include a **reasoning-level indicator** in the system prompt using the `Reasoning-level:` directive. The available levels are:
 
 
40
 
41
+ | Level | Behavior |
42
+ |----------|--------|
43
+ | `none` | No reasoning trace generated; direct response only. |
44
+ | `low` | Minimal internal reasoning; concise output. |
45
+ | `medium` | Balanced reasoning with intermediate steps. |
46
+ | `high` | Detailed step-by-step reasoning with full justification. |
47
+ | `ultra` | Maximal depth reasoning; exhaustive exploration of alternatives and edge cases. |
48
 
49
+ > **Note**: The absence of a valid reasoning level will result in undefined behavior. Always specify one.
 
50
 
51
+ ### Example Prompt (ChatML Format):
 
 
 
 
 
52
 
53
+ ```xml
54
+ <|im_start|>system
55
+ Reasoning-level: high
56
+ <|im_end|>
57
+ <|im_start|>user
58
+ What is the capital of France?
59
+ <|im_end|>
60
+ <|im_start|>assistant
61
+ <think>
62
+ To determine the capital of France, we recall that France is a country in Western Europe. Its major cities include Paris, Lyon, Marseille, and Toulouse. Among these, Paris is historically and politically recognized as the capital. It hosts the French government, national institutions, and the President’s residence. Therefore, the capital of France is Paris.
63
+ </think>
64
+ Paris
65
  ```
66
 
67
+ ---
68
 
69
+ ## Important Notes
70
+
71
+ - This is a **preview release**. Performance may vary across edge cases or non-standard inputs.
72
+ - For production applications, please wait for the official release and corresponding documentation.
73
+ - Model outputs should be validated for accuracy, especially in safety-critical domains.
74
+
75
+ ---