squ11z1 commited on
Commit
93458f7
·
verified ·
1 Parent(s): 30494aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -6
README.md CHANGED
@@ -1,10 +1,66 @@
1
  ---
2
- title: README
3
- emoji: 👀
4
- colorFrom: pink
5
- colorTo: red
6
  sdk: static
7
- pinned: false
 
 
8
  ---
9
 
10
- Edit this `README.md` markdown file to author your organization card.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: MerlinResearch
3
+ emoji: 🛡️
4
+ colorFrom: purple
5
+ colorTo: purple
6
  sdk: static
7
+ pinned: true
8
+ license: apache-2.0
9
+ short_description: AI safety, reasoning, and alignment research lab.
10
  ---
11
 
12
+ # MerlinSafety
13
+
14
+ **MerlinSafety** is an independent AI safety and reasoning research organization focused on building practical, auditable, and robust open models.
15
+
16
+ ## Mission
17
+
18
+ We develop and evaluate models that are:
19
+ - Strong in constrained instruction-following
20
+ - Safer in real-world agentic workflows
21
+ - Better aligned under uncertainty and adversarial prompts
22
+ - Transparent in behavior, limits, and deployment risks
23
+
24
+ ## What We Build
25
+
26
+ - Safety-oriented reasoning models
27
+ - Alignment-focused post-training pipelines
28
+ - Evaluation suites for robustness, controllability, and failure analysis
29
+ - Open artifacts for reproducible research
30
+
31
+ ## Current Focus Areas
32
+
33
+ - Safety reasoning for small/efficient LLMs
34
+ - Misalignment reduction via structured post-training
35
+ - Hallucination risk reduction in high-stakes contexts
36
+ - Robust instruction adherence with explicit constraints
37
+
38
+ ## Research Principles
39
+
40
+ 1. **Measure behavior, not marketing claims.**
41
+ 2. **Prioritize reproducibility and clear documentation.**
42
+ 3. **Publish limitations, not only strengths.**
43
+ 4. **Design for safe deployment from day one.**
44
+
45
+ ## Models
46
+
47
+ Our flagship releases are published under this organization with:
48
+ - Full model cards
49
+ - Clear training/deployment notes
50
+ - Practical usage guidance
51
+
52
+ ## Collaboration
53
+
54
+ We welcome collaboration on:
55
+ - AI safety evaluation
56
+ - Alignment methods
57
+ - Reasoning benchmarks
58
+ - Responsible open model deployment
59
+
60
+ For partnerships or research collaboration, contact us via Hugging Face discussions or linked channels in our repositories.
61
+
62
+ ---
63
+
64
+ **MerlinSafety**
65
+ Safe reasoning. Measurable alignment. Real-world robustness.
66
+