|
|
--- |
|
|
base_model: |
|
|
- ByteDance-Seed/Seed-Coder-8B-Reasoning |
|
|
tags: |
|
|
- text-generation-inference |
|
|
- transformers |
|
|
- unsloth |
|
|
- llama |
|
|
license: mit |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
 |
|
|
|
|
|
# Daedalus-1-8B |
|
|
|
|
|
[](https://huggingface.co/NoemaResearch/Daedalus-1-8B) |
|
|
[](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning) |
|
|
[](LICENSE) |
|
|
|
|
|
Daedalus-1-8B is an 8 billion parameter language model for code generation and reasoning, developed by **Noema Research**. |
|
|
It is a finetuned derivative of [Seed-Coder-8B-Reasoning](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning), |
|
|
with enhancements for instruction following, structured code generation, and improved safety alignment. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Overview |
|
|
|
|
|
- **Base model:** `ByteDance-Seed/Seed-Coder-8B-Reasoning` |
|
|
- **Architecture:** Decoder-only transformer |
|
|
- **Parameters:** ~8.25B |
|
|
- **Context length:** Long-context support (up to ~64k tokens) |
|
|
- **Domain:** Programming and natural language reasoning |
|
|
- **Primary applications:** |
|
|
- Code generation and completion |
|
|
- Debugging and error explanation |
|
|
- Unit test generation |
|
|
- Structured outputs (e.g., JSON, function calls) |
|
|
- **License:** MIT |
|
|
|
|
|
--- |
|
|
|
|
|
## Key Improvements |
|
|
|
|
|
Relative to the base model, Daedalus introduces targeted post-training improvements: |
|
|
|
|
|
- **Instruction tuning** for developer-oriented tasks |
|
|
- **Structured output fidelity**, supporting JSON and schema-constrained responses |
|
|
- **Enhanced reasoning** for debugging and multi-step problem solving |
|
|
- **Reduced error rate** in code execution benchmarks |
|
|
- **Safety-oriented adjustments**, including avoidance of unsafe coding patterns |
|
|
|
|
|
--- |
|
|
|
|
|
## Usage |
|
|
|
|
|
The model is released in Hugging Face Transformers format. Example: |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
import torch |
|
|
|
|
|
model_id = "NoemaResearch/Daedalus-1-8B" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_id, |
|
|
torch_dtype=torch.bfloat16, |
|
|
device_map="auto", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
messages = [ |
|
|
{"role":"system", "content":"You are Daedalus, a coding assistant."}, |
|
|
{"role":"user", "content":"Write a memory-efficient quicksort in Python with unit tests."} |
|
|
] |
|
|
|
|
|
inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device) |
|
|
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.2, top_p=0.95) |
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
|
```` |
|
|
|
|
|
**Recommended settings:** |
|
|
|
|
|
* `temperature=0.2–0.6` for deterministic code generation |
|
|
* `top_p=0.9–0.95` for balanced creativity and correctness |
|
|
|
|
|
--- |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
Daedalus inherits strong performance on competitive programming and reasoning tasks from Seed-Coder-8B-Reasoning. |
|
|
Internal evaluations indicate: |
|
|
|
|
|
* Higher **unit test pass rates** |
|
|
* Improved **structured output validity** |
|
|
* Reduced incidence of **hallucinated APIs** |
|
|
|
|
|
A comprehensive benchmark report will be released in future updates. |
|
|
For upstream benchmarks, please refer to the [Seed-Coder-8B-Reasoning model card](https://huggingface.co/ByteDance-Seed/Seed-Coder-8B-Reasoning). |
|
|
|
|
|
--- |
|
|
|
|
|
## Limitations |
|
|
|
|
|
Daedalus remains subject to common limitations of large language models: |
|
|
|
|
|
* **Hallucinated libraries or functions:** the model may generate non-existent APIs |
|
|
* **Insecure coding patterns:** suggestions should be reviewed for security and safety |
|
|
* **Reasoning errors:** multi-step solutions may fail on complex edge cases |
|
|
* **Dependence on prompt quality:** outputs are sensitive to phrasing and context |
|
|
|
|
|
All generated code should be verified, linted, and tested before use in production. |
|
|
|
|
|
--- |
|
|
|
|
|
## Responsible Use |
|
|
|
|
|
* Do not provide secrets or credentials in prompts. |
|
|
* Use outputs only in controlled, sandboxed, or reviewed environments. |
|
|
* The model should not be employed for generating malicious software or unsafe code. |
|
|
* We encourage the use of additional guardrails (static analyzers, test harnesses, execution sandboxes) in deployment contexts. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Variants |
|
|
|
|
|
* **Full-precision (safetensors)** — for research and high-fidelity inference |
|
|
* **bf16 / fp16** — for efficient inference on modern accelerators |
|
|
* **Quantized variants (int8, int4)** — for resource-constrained environments |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this model, please cite both Daedalus and the underlying Seed-Coder base model: |
|
|
|
|
|
```bibtex |
|
|
@misc{noema2025daedalus, |
|
|
title={Daedalus-1-8B}, |
|
|
author={Noema Research}, |
|
|
year={2025}, |
|
|
howpublished={\url{https://huggingface.co/NoemaResearch/Daedalus-1-8B}} |
|
|
} |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## Acknowledgements |
|
|
|
|
|
Daedalus builds upon the [Seed-Coder](https://huggingface.co/ByteDance-Seed) family of models developed by ByteDance-Seed. |
|
|
We thank the Seed team for releasing their models under permissive terms, enabling further research and refinement. |