Spestly commited on
Commit
b45420d
·
verified ·
1 Parent(s): 7ba6ee9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +142 -2
README.md CHANGED
@@ -1,11 +1,151 @@
1
  ---
2
- base_model: ByteDance-Seed/Seed-Coder-8B-Reasoning-bf16
 
3
  tags:
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - llama
8
- license: apache-2.0
9
  language:
10
  - en
11
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - ByteDance-Seed/Seed-Coder-8B-Reasoning
4
  tags:
5
  - text-generation-inference
6
  - transformers
7
  - unsloth
8
  - llama
9
+ license: mit
10
  language:
11
  - en
12
  ---
13
+
14
+ # Daedalus-1-8B
15
+
16
+ [![Model](https://img.shields.io/badge/Model-Daedalus--1--8B-blue)](https://huggingface.co/NoemaResearch/Daedalus-1-8B)
17
+ [![Base](https://img.shields.io/badge/Base-Seed--Coder--8B--Reasoning-green)](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning)
18
+ [![License](https://img.shields.io/badge/License-MIT-yellow)](LICENSE)
19
+
20
+ Daedalus-1-8B is an 8 billion parameter language model for code generation and reasoning, developed by **Noema Research**.
21
+ It is a finetuned derivative of [Seed-Coder-8B-Reasoning](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning),
22
+ with enhancements for instruction following, structured code generation, and improved safety alignment.
23
+
24
+ ---
25
+
26
+ ## Model Overview
27
+
28
+ - **Base model:** `ByteDance-Seed/Seed-Coder-8B-Reasoning`
29
+ - **Architecture:** Decoder-only transformer
30
+ - **Parameters:** ~8.25B
31
+ - **Context length:** Long-context support (up to ~64k tokens)
32
+ - **Domain:** Programming and natural language reasoning
33
+ - **Primary applications:**
34
+ - Code generation and completion
35
+ - Debugging and error explanation
36
+ - Unit test generation
37
+ - Structured outputs (e.g., JSON, function calls)
38
+ - **License:** MIT
39
+
40
+ ---
41
+
42
+ ## Key Improvements
43
+
44
+ Relative to the base model, Daedalus introduces targeted post-training improvements:
45
+
46
+ - **Instruction tuning** for developer-oriented tasks
47
+ - **Structured output fidelity**, supporting JSON and schema-constrained responses
48
+ - **Enhanced reasoning** for debugging and multi-step problem solving
49
+ - **Reduced error rate** in code execution benchmarks
50
+ - **Safety-oriented adjustments**, including avoidance of unsafe coding patterns
51
+
52
+ ---
53
+
54
+ ## Usage
55
+
56
+ The model is released in Hugging Face Transformers format. Example:
57
+
58
+ ```python
59
+ from transformers import AutoTokenizer, AutoModelForCausalLM
60
+ import torch
61
+
62
+ model_id = "NoemaResearch/Daedalus-1-8B"
63
+
64
+ tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
65
+ model = AutoModelForCausalLM.from_pretrained(
66
+ model_id,
67
+ torch_dtype=torch.bfloat16,
68
+ device_map="auto",
69
+ trust_remote_code=True
70
+ )
71
+
72
+ messages = [
73
+ {"role":"system", "content":"You are Daedalus, a coding assistant."},
74
+ {"role":"user", "content":"Write a memory-efficient quicksort in Python with unit tests."}
75
+ ]
76
+
77
+ inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
78
+ outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.2, top_p=0.95)
79
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
80
+ ````
81
+
82
+ **Recommended settings:**
83
+
84
+ * `temperature=0.2–0.6` for deterministic code generation
85
+ * `top_p=0.9–0.95` for balanced creativity and correctness
86
+
87
+ ---
88
+
89
+ ## Evaluation
90
+
91
+ Daedalus inherits strong performance on competitive programming and reasoning tasks from Seed-Coder-8B-Reasoning.
92
+ Internal evaluations indicate:
93
+
94
+ * Higher **unit test pass rates**
95
+ * Improved **structured output validity**
96
+ * Reduced incidence of **hallucinated APIs**
97
+
98
+ A comprehensive benchmark report will be released in future updates.
99
+ For upstream benchmarks, please refer to the [Seed-Coder-8B-Reasoning model card](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning).
100
+
101
+ ---
102
+
103
+ ## Limitations
104
+
105
+ Daedalus remains subject to common limitations of large language models:
106
+
107
+ * **Hallucinated libraries or functions:** the model may generate non-existent APIs
108
+ * **Insecure coding patterns:** suggestions should be reviewed for security and safety
109
+ * **Reasoning errors:** multi-step solutions may fail on complex edge cases
110
+ * **Dependence on prompt quality:** outputs are sensitive to phrasing and context
111
+
112
+ All generated code should be verified, linted, and tested before use in production.
113
+
114
+ ---
115
+
116
+ ## Responsible Use
117
+
118
+ * Do not provide secrets or credentials in prompts.
119
+ * Use outputs only in controlled, sandboxed, or reviewed environments.
120
+ * The model should not be employed for generating malicious software or unsafe code.
121
+ * We encourage the use of additional guardrails (static analyzers, test harnesses, execution sandboxes) in deployment contexts.
122
+
123
+ ---
124
+
125
+ ## Model Variants
126
+
127
+ * **Full-precision (safetensors)** — for research and high-fidelity inference
128
+ * **bf16 / fp16** — for efficient inference on modern accelerators
129
+ * **Quantized variants (int8, int4)** — for resource-constrained environments
130
+
131
+ ---
132
+
133
+ ## Citation
134
+
135
+ If you use this model, please cite both Daedalus and the underlying Seed-Coder base model:
136
+
137
+ ```bibtex
138
+ @misc{noema2025daedalus,
139
+ title={Daedalus-1-8B},
140
+ author={Noema Research},
141
+ year={2025},
142
+ howpublished={\url{https://huggingface.co/NoemaResearch/Daedalus-1-8B}}
143
+ }
144
+ ```
145
+
146
+ ---
147
+
148
+ ## Acknowledgements
149
+
150
+ Daedalus builds upon the [Seed-Coder](https://huggingface.co/ByteDance-Seed) family of models developed by ByteDance-Seed.
151
+ We thank the Seed team for releasing their models under permissive terms, enabling further research and refinement.