Update README.md
Browse files
README.md
CHANGED
|
@@ -1,11 +1,151 @@
|
|
| 1 |
---
|
| 2 |
-
base_model:
|
|
|
|
| 3 |
tags:
|
| 4 |
- text-generation-inference
|
| 5 |
- transformers
|
| 6 |
- unsloth
|
| 7 |
- llama
|
| 8 |
-
license:
|
| 9 |
language:
|
| 10 |
- en
|
| 11 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- ByteDance-Seed/Seed-Coder-8B-Reasoning
|
| 4 |
tags:
|
| 5 |
- text-generation-inference
|
| 6 |
- transformers
|
| 7 |
- unsloth
|
| 8 |
- llama
|
| 9 |
+
license: mit
|
| 10 |
language:
|
| 11 |
- en
|
| 12 |
---
|
| 13 |
+
|
| 14 |
+
# Daedalus-1-8B
|
| 15 |
+
|
| 16 |
+
[](https://huggingface.co/NoemaResearch/Daedalus-1-8B)
|
| 17 |
+
[](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning)
|
| 18 |
+
[](LICENSE)
|
| 19 |
+
|
| 20 |
+
Daedalus-1-8B is an 8 billion parameter language model for code generation and reasoning, developed by **Noema Research**.
|
| 21 |
+
It is a finetuned derivative of [Seed-Coder-8B-Reasoning](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning),
|
| 22 |
+
with enhancements for instruction following, structured code generation, and improved safety alignment.
|
| 23 |
+
|
| 24 |
+
---
|
| 25 |
+
|
| 26 |
+
## Model Overview
|
| 27 |
+
|
| 28 |
+
- **Base model:** `ByteDance-Seed/Seed-Coder-8B-Reasoning`
|
| 29 |
+
- **Architecture:** Decoder-only transformer
|
| 30 |
+
- **Parameters:** ~8.25B
|
| 31 |
+
- **Context length:** Long-context support (up to ~64k tokens)
|
| 32 |
+
- **Domain:** Programming and natural language reasoning
|
| 33 |
+
- **Primary applications:**
|
| 34 |
+
- Code generation and completion
|
| 35 |
+
- Debugging and error explanation
|
| 36 |
+
- Unit test generation
|
| 37 |
+
- Structured outputs (e.g., JSON, function calls)
|
| 38 |
+
- **License:** MIT
|
| 39 |
+
|
| 40 |
+
---
|
| 41 |
+
|
| 42 |
+
## Key Improvements
|
| 43 |
+
|
| 44 |
+
Relative to the base model, Daedalus introduces targeted post-training improvements:
|
| 45 |
+
|
| 46 |
+
- **Instruction tuning** for developer-oriented tasks
|
| 47 |
+
- **Structured output fidelity**, supporting JSON and schema-constrained responses
|
| 48 |
+
- **Enhanced reasoning** for debugging and multi-step problem solving
|
| 49 |
+
- **Reduced error rate** in code execution benchmarks
|
| 50 |
+
- **Safety-oriented adjustments**, including avoidance of unsafe coding patterns
|
| 51 |
+
|
| 52 |
+
---
|
| 53 |
+
|
| 54 |
+
## Usage
|
| 55 |
+
|
| 56 |
+
The model is released in Hugging Face Transformers format. Example:
|
| 57 |
+
|
| 58 |
+
```python
|
| 59 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 60 |
+
import torch
|
| 61 |
+
|
| 62 |
+
model_id = "NoemaResearch/Daedalus-1-8B"
|
| 63 |
+
|
| 64 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
| 65 |
+
model = AutoModelForCausalLM.from_pretrained(
|
| 66 |
+
model_id,
|
| 67 |
+
torch_dtype=torch.bfloat16,
|
| 68 |
+
device_map="auto",
|
| 69 |
+
trust_remote_code=True
|
| 70 |
+
)
|
| 71 |
+
|
| 72 |
+
messages = [
|
| 73 |
+
{"role":"system", "content":"You are Daedalus, a coding assistant."},
|
| 74 |
+
{"role":"user", "content":"Write a memory-efficient quicksort in Python with unit tests."}
|
| 75 |
+
]
|
| 76 |
+
|
| 77 |
+
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
| 78 |
+
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.2, top_p=0.95)
|
| 79 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 80 |
+
````
|
| 81 |
+
|
| 82 |
+
**Recommended settings:**
|
| 83 |
+
|
| 84 |
+
* `temperature=0.2–0.6` for deterministic code generation
|
| 85 |
+
* `top_p=0.9–0.95` for balanced creativity and correctness
|
| 86 |
+
|
| 87 |
+
---
|
| 88 |
+
|
| 89 |
+
## Evaluation
|
| 90 |
+
|
| 91 |
+
Daedalus inherits strong performance on competitive programming and reasoning tasks from Seed-Coder-8B-Reasoning.
|
| 92 |
+
Internal evaluations indicate:
|
| 93 |
+
|
| 94 |
+
* Higher **unit test pass rates**
|
| 95 |
+
* Improved **structured output validity**
|
| 96 |
+
* Reduced incidence of **hallucinated APIs**
|
| 97 |
+
|
| 98 |
+
A comprehensive benchmark report will be released in future updates.
|
| 99 |
+
For upstream benchmarks, please refer to the [Seed-Coder-8B-Reasoning model card](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning).
|
| 100 |
+
|
| 101 |
+
---
|
| 102 |
+
|
| 103 |
+
## Limitations
|
| 104 |
+
|
| 105 |
+
Daedalus remains subject to common limitations of large language models:
|
| 106 |
+
|
| 107 |
+
* **Hallucinated libraries or functions:** the model may generate non-existent APIs
|
| 108 |
+
* **Insecure coding patterns:** suggestions should be reviewed for security and safety
|
| 109 |
+
* **Reasoning errors:** multi-step solutions may fail on complex edge cases
|
| 110 |
+
* **Dependence on prompt quality:** outputs are sensitive to phrasing and context
|
| 111 |
+
|
| 112 |
+
All generated code should be verified, linted, and tested before use in production.
|
| 113 |
+
|
| 114 |
+
---
|
| 115 |
+
|
| 116 |
+
## Responsible Use
|
| 117 |
+
|
| 118 |
+
* Do not provide secrets or credentials in prompts.
|
| 119 |
+
* Use outputs only in controlled, sandboxed, or reviewed environments.
|
| 120 |
+
* The model should not be employed for generating malicious software or unsafe code.
|
| 121 |
+
* We encourage the use of additional guardrails (static analyzers, test harnesses, execution sandboxes) in deployment contexts.
|
| 122 |
+
|
| 123 |
+
---
|
| 124 |
+
|
| 125 |
+
## Model Variants
|
| 126 |
+
|
| 127 |
+
* **Full-precision (safetensors)** — for research and high-fidelity inference
|
| 128 |
+
* **bf16 / fp16** — for efficient inference on modern accelerators
|
| 129 |
+
* **Quantized variants (int8, int4)** — for resource-constrained environments
|
| 130 |
+
|
| 131 |
+
---
|
| 132 |
+
|
| 133 |
+
## Citation
|
| 134 |
+
|
| 135 |
+
If you use this model, please cite both Daedalus and the underlying Seed-Coder base model:
|
| 136 |
+
|
| 137 |
+
```bibtex
|
| 138 |
+
@misc{noema2025daedalus,
|
| 139 |
+
title={Daedalus-1-8B},
|
| 140 |
+
author={Noema Research},
|
| 141 |
+
year={2025},
|
| 142 |
+
howpublished={\url{https://huggingface.co/NoemaResearch/Daedalus-1-8B}}
|
| 143 |
+
}
|
| 144 |
+
```
|
| 145 |
+
|
| 146 |
+
---
|
| 147 |
+
|
| 148 |
+
## Acknowledgements
|
| 149 |
+
|
| 150 |
+
Daedalus builds upon the [Seed-Coder](https://huggingface.co/ByteDance-Seed) family of models developed by ByteDance-Seed.
|
| 151 |
+
We thank the Seed team for releasing their models under permissive terms, enabling further research and refinement.
|