Running on Zero MCP Featured 116 FLUX.2 Klein 9B KV π¨ 116 Generate or edit images from text and optional photos
view post Post 5384 Surya-1.1T: Scaling Beyond Human-Level Reasoning via 146 Trillion Token Pre-trainingAuthor: SKT AI LABSAffiliation: SKT AI Labs / Project SuryaModel Architecture: Optimized Dense TransformerParameters: 1.1 TrillionTraining Tokens: 146 TrillionWanna collaborate us Friends let's Start Journey we have Collected 146 trillon tokens and done pre training but we need to made more powerfullWhitepaper - https://github.com/SHRIJANAGAIN/PROFF See translation 50 replies Β· π₯ 15 15 π 9 9 π 8 8 π€ 7 7 β 7 7 π 6 6 β€οΈ 6 6 π 5 5 π§ 5 5 π€ 5 5 π€― 3 3 + Reply
view post Post 3820 Can small models program?Although even if they are reasoning AIs, small AIs cannot create extensive and high-quality code, at least that's what is commonly thought.We present OrionLLM/NanoCoder-0.6b, an AI with just 600 million parameters based on qwen3-0.6b and trained with the dataset nvidia/OpenCodeReasoning.While not good at complex code, we observed a significant improvement in code generation (especially in Python code), demonstrating that, when trained correctly, small AIs can, in fact, program. See translation 2 replies Β· π€ 9 9 π 1 1 π₯ 1 1 + Reply