Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,10 +1,66 @@
|
|
| 1 |
---
|
| 2 |
-
title:
|
| 3 |
-
emoji:
|
| 4 |
-
colorFrom:
|
| 5 |
-
colorTo:
|
| 6 |
sdk: static
|
| 7 |
-
pinned:
|
|
|
|
|
|
|
| 8 |
---
|
| 9 |
|
| 10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
title: MerlinResearch
|
| 3 |
+
emoji: 🛡️
|
| 4 |
+
colorFrom: purple
|
| 5 |
+
colorTo: purple
|
| 6 |
sdk: static
|
| 7 |
+
pinned: true
|
| 8 |
+
license: apache-2.0
|
| 9 |
+
short_description: AI safety, reasoning, and alignment research lab.
|
| 10 |
---
|
| 11 |
|
| 12 |
+
# MerlinSafety
|
| 13 |
+
|
| 14 |
+
**MerlinSafety** is an independent AI safety and reasoning research organization focused on building practical, auditable, and robust open models.
|
| 15 |
+
|
| 16 |
+
## Mission
|
| 17 |
+
|
| 18 |
+
We develop and evaluate models that are:
|
| 19 |
+
- Strong in constrained instruction-following
|
| 20 |
+
- Safer in real-world agentic workflows
|
| 21 |
+
- Better aligned under uncertainty and adversarial prompts
|
| 22 |
+
- Transparent in behavior, limits, and deployment risks
|
| 23 |
+
|
| 24 |
+
## What We Build
|
| 25 |
+
|
| 26 |
+
- Safety-oriented reasoning models
|
| 27 |
+
- Alignment-focused post-training pipelines
|
| 28 |
+
- Evaluation suites for robustness, controllability, and failure analysis
|
| 29 |
+
- Open artifacts for reproducible research
|
| 30 |
+
|
| 31 |
+
## Current Focus Areas
|
| 32 |
+
|
| 33 |
+
- Safety reasoning for small/efficient LLMs
|
| 34 |
+
- Misalignment reduction via structured post-training
|
| 35 |
+
- Hallucination risk reduction in high-stakes contexts
|
| 36 |
+
- Robust instruction adherence with explicit constraints
|
| 37 |
+
|
| 38 |
+
## Research Principles
|
| 39 |
+
|
| 40 |
+
1. **Measure behavior, not marketing claims.**
|
| 41 |
+
2. **Prioritize reproducibility and clear documentation.**
|
| 42 |
+
3. **Publish limitations, not only strengths.**
|
| 43 |
+
4. **Design for safe deployment from day one.**
|
| 44 |
+
|
| 45 |
+
## Models
|
| 46 |
+
|
| 47 |
+
Our flagship releases are published under this organization with:
|
| 48 |
+
- Full model cards
|
| 49 |
+
- Clear training/deployment notes
|
| 50 |
+
- Practical usage guidance
|
| 51 |
+
|
| 52 |
+
## Collaboration
|
| 53 |
+
|
| 54 |
+
We welcome collaboration on:
|
| 55 |
+
- AI safety evaluation
|
| 56 |
+
- Alignment methods
|
| 57 |
+
- Reasoning benchmarks
|
| 58 |
+
- Responsible open model deployment
|
| 59 |
+
|
| 60 |
+
For partnerships or research collaboration, contact us via Hugging Face discussions or linked channels in our repositories.
|
| 61 |
+
|
| 62 |
+
---
|
| 63 |
+
|
| 64 |
+
**MerlinSafety**
|
| 65 |
+
Safe reasoning. Measurable alignment. Real-world robustness.
|
| 66 |
+
|