#AI2 releases OLMo 2 32B, trained on 6T tokens with #Tulu3.1 post-training. Matches or exceeds GPT3.5 Turbo while using just 1/3 the compute of #Qwen2.5 32B. Complete open recipe includes data, code, weights and training methodology.
๐๐ข๐'๐ฌ ๐๐๐ฐ ๐๐ฉ๐๐ง-๐๐จ๐ฎ๐ซ๐๐ ๐๐จ๐๐๐ฅ ๐๐ฎ๐ฅ๐ฎ ๐ ๐๐๐๐ ๐๐ฎ๐ญ๐ฉ๐๐ซ๐๐จ๐ซ๐ฆ๐ฌ ๐๐๐๐ฉ๐๐๐๐ค ๐๐ง๐ ๐๐๐-๐๐จ
With 405B parameters and post-training using RLVR, itโs a major breakthrough. Open-source, accessible, and efficient, it tops PopQA, GSM8K, and more! โก๐ง
tinyurl.com/vjk6kp8x
#AI #Tulu3 #MachineLearning #OpenSource
๐ Ai2'dan DeepSeek'e meydan okuyan yeni AI modeli!
ABD merkezli Ai2, aรงฤฑk kaynaklฤฑ Tulu3-405B modelinin DeepSeek V3 ve bazฤฑ testlerde GPT-4oโyu geรงtiฤini duyurdu. 405 milyar parametreye sahip model, AI dรผnyasฤฑnda yeni bir dรถnรผm noktasฤฑ olabilir.
#Teknoloji #Haber #YapayZeka #Tulu3 #DeepSeek
Exciting news! Ai2 has just released Tรผlu 3, a groundbreaking open-source AI model that outperforms DeepSeek v3 and GPT-4o on key benchmarks. With 405 billion parameters, Tรผlu 3 is set to revolutionize the AI landscape. #AI #OpenSource #Tulu3 #DeepSeek #TechNews
youtube.com/shorts/UQfNf...
TรLU 3 Pushes the Boundaries of AI Post-Training Excellence ๐ฌโจ๐ www.azoai.com/news/2024120... #AI #MachineLearning #OpenSource #LanguageModels #PostTraining #TULU3 #Innovation #TechResearch #RLVR @alleninstitute.bsky.social