huu-ontocord commited on
Commit
1c2c7ce
·
verified ·
1 Parent(s): c7bba1c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -1
README.md CHANGED
@@ -6,4 +6,11 @@ colorTo: yellow
6
  sdk: streamlit
7
  pinned: false
8
  ---
9
- We are a group of volunteer researchers focused on equal access to multilingual AI. We present a set of red-teamed models. Trained on the LUMI HPC in Finland (thus the name Aurora). The "-m" designation stands for multimodal, multilingual, multidomain mixture of expert (MOE) models, each of which we intend to research. As part of Ontocord.AI's dedication to lawful open science AI, Ontocord coordinated to this effort with the volunteers and contributed to the safety measures. This work should NOT be confused with the AuroraGPT, https://www.hpcwire.com/2023/11/13/training-of-1-trillion-parameter-scientific-ai-begins/.
 
 
 
 
 
 
 
 
6
  sdk: streamlit
7
  pinned: false
8
  ---
9
+ We are a group of volunteer researchers dedicated to promoting equal access to multimodal and multilingual AI. Our goal is to build a permissive and open stack for developing multimodal LLMs. This initiative is a collaborative effort led by OntocordAI. We began as an effort named MDEL (Multi-Domain Expert Learning).
10
+
11
+ The -m in Aurora-M2 refers to our focus on multimodal, multilingual, multidomain mixture-of-experts (MoE) models, each of which we aim to explore and develop through ongoing research.
12
+
13
+ Building on our previous success— (Aurora-M: Open Source Continual Pre-training for Multilingual Language and Code)[https://aclanthology.org/2025.coling-industry.56/] — we are training a family of models aligned with laws, regulations, and policies for controllable AI. The series will include models with parameter sizes of 3B, 8B, and 21B, aligned with the comprehensive policy framework of the EU AI Act, specifically Annex III of the Act.
14
+
15
+ As part of our commitment to openness, we plan to open-source the entire training pipeline and experimental process—including data synthesis and the evolving methodologies we employ in model training. Stay with us!
16
+