view post Post 5420 Surya-1.1T: Scaling Beyond Human-Level Reasoning via 146 Trillion Token Pre-trainingAuthor: SKT AI LABSAffiliation: SKT AI Labs / Project SuryaModel Architecture: Optimized Dense TransformerParameters: 1.1 TrillionTraining Tokens: 146 TrillionWanna collaborate us Friends let's Start Journey we have Collected 146 trillon tokens and done pre training but we need to made more powerfullWhitepaper - https://github.com/SHRIJANAGAIN/PROFF See translation 56 replies · 🔥 15 15 👍 9 9 🚀 8 8 🤗 7 7 ➕ 7 7 👀 6 6 ❤️ 6 6 😎 5 5 🧠 5 5 🤝 5 5 🤯 3 3 + Reply
Andycurrent/Gemma-3-4B-VL-it-Gemini-Pro-Heretic-Uncensored-Thinking_GGUF Image-Text-to-Text • 4B • Updated 24 days ago • 42.4k • 45