view post Post 5356 Surya-1.1T: Scaling Beyond Human-Level Reasoning via 146 Trillion Token Pre-trainingAuthor: SKT AI LABSAffiliation: SKT AI Labs / Project SuryaModel Architecture: Optimized Dense TransformerParameters: 1.1 TrillionTraining Tokens: 146 TrillionWanna collaborate us Friends let's Start Journey we have Collected 146 trillon tokens and done pre training but we need to made more powerfullWhitepaper - https://github.com/SHRIJANAGAIN/PROFF See translation 49 replies · 🔥 15 15 👍 9 9 🚀 8 8 🤗 7 7 ➕ 7 7 👀 6 6 ❤️ 6 6 😎 5 5 🧠 5 5 🤝 5 5 🤯 3 3 + Reply
Quantifying the Carbon Emissions of Machine Learning Paper • 1910.09700 • Published Oct 21, 2019 • 39