rasdani_'s profile picture. hill climbing @PrimeIntellect

rasdani

@rasdani_

hill climbing @PrimeIntellect

مثبتة

agi was the friends we made along the way

rasdani_'s tweet image. agi was the friends we made along the way

Introducing INTELLECT-3: Scaling RL to a 100B+ MoE model on our end-to-end stack Achieving state-of-the-art performance for its size across math, code and reasoning Built using the same tools we put in your hands, from environments & evals, RL frameworks, sandboxes & more



rasdani أعاد

should've shot farther :^)

≈320B 15B active, calling it Trillion-class on the next turn go on, surprise me



rasdani أعاد

Details on the trinity architecture that we settled on!

stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!

rasdani أعاد

> be arcee > look around > realize open-weight frontier MoE is basically a Qwen/DeepSeek monopoly > decide “nah, we’re building our own” > actual end-to-end pretraining > on US soil > introducing Trinity > Nano (6B MoE) and Mini (26B MoE) > open weights, Apache 2.0 > free on…

TheAhmadOsman's tweet image. > be arcee
> look around
> realize open-weight frontier MoE is basically a Qwen/DeepSeek monopoly
> decide “nah, we’re building our own”
> actual end-to-end pretraining
> on US soil

> introducing Trinity
> Nano (6B MoE) and Mini (26B MoE)
> open weights, Apache 2.0
> free on…

rasdani أعاد

few people exemplify the spirit of "just doing things" to the degree that @stochasticchasm does. dude is 25 and has been at @arcee_ai less than a year and is driving pretraining efforts which rival the ambition of pretty much anything else in the open. this is a moneyball moment

Details on the trinity architecture that we settled on!

stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!
stochasticchasm's tweet image. Details on the trinity architecture that we settled on!


rasdani أعاد

ArceeAI Just dropped Trinity models -mini and -nano variants with sizes 26B and 6B, there is a 3rd one with larger size (420B !)cooking still. Read the details in blog arcee.ai/blog/the-trini…


rasdani أعاد

Hell of a day to launch

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.



rasdani أعاد

This is how you train SOTA models in late 2025.

Dorialexander's tweet image. This is how you train SOTA models in late 2025.

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.



rasdani أعاد

The 420B-A13B is coming in January which will be SO BIG! The biggest American pretrained model released in a long time

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.



rasdani أعاد

cant wait to post train this model family

We're excited to support @Arcee_ai's Trinity models A family of open base Mixture of Experts models pretrained in collaboration between Arcee, Datology, and Prime Intellect Releasing Trinity Nano (6B) and Mini (26B) today, with Trinity Large still in training



rasdani أعاد

it's the dawn of a glorious new era for western open models congrats @arcee_ai on an incredible release!! thrilled we got to be a part of it :)

willccbb's tweet image. it's the dawn of a glorious new era for western open models

congrats @arcee_ai on an incredible release!! thrilled we got to be a part of it :)
willccbb's tweet image. it's the dawn of a glorious new era for western open models

congrats @arcee_ai on an incredible release!! thrilled we got to be a part of it :)
willccbb's tweet image. it's the dawn of a glorious new era for western open models

congrats @arcee_ai on an incredible release!! thrilled we got to be a part of it :)

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.



rasdani أعاد

Full thread with blog post and benchmark results: x.com/latkins/status…

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.



rasdani أعاد

Arcee's Trinity Mini and Nano are live now! It's been an incredible couple of months collaborating on this pre-training run with the great folks at @arcee_ai and @datologyai. Job's not finished. Trinity Large is still training 🫡

We're excited to support @Arcee_ai's Trinity models A family of open base Mixture of Experts models pretrained in collaboration between Arcee, Datology, and Prime Intellect Releasing Trinity Nano (6B) and Mini (26B) today, with Trinity Large still in training



rasdani أعاد

We're excited to support @Arcee_ai's Trinity models A family of open base Mixture of Experts models pretrained in collaboration between Arcee, Datology, and Prime Intellect Releasing Trinity Nano (6B) and Mini (26B) today, with Trinity Large still in training


rasdani أعاد

Today, we are introducing Trinity, the start of an open-weight MoE family that businesses and developers can own. Trinity-Mini (26B-A3B) Trinity-Nano-Preview (6B-A1B) Available Today on Huggingface.


rasdani أعاد

you can now run prime-rl on kubernetes via helm


rasdani أعاد

This is what I have using prime's exact settings: aime2024: 89.79% aime2025: 87.71% avg@32 w/ temp=0.6, top_p=1.0, max_tokens=81920 template = "Solve the following math problem. Explain your reasoning and put the final answer in \\boxed{{}}.\n\n{prompt}"


rasdani أعاد

I’ve been able to match prime’s results for INTELLECT-3 closely, within 1% of aime scores reported, using longer response length limits and the exact prompt they used for evaluation. (There was some clipping at 65k.) GLM-4.5-Air is still running, will update when it lands :)

Adding my local results to the discussion. I haven't been able to reproduce the claimed 5%+ gap in prime's headline plots, and I'm seeing a difference within noise with several different prompt formats. x.com/PrimeIntellect…

rawsh0's tweet image. Adding my local results to the discussion. I haven't been able to reproduce the claimed 5%+ gap in prime's headline plots, and I'm seeing a difference within noise with several different prompt formats.
x.com/PrimeIntellect…


rasdani أعاد

What's really interesting here is that Prime Intellect has quietly built one of the best post-training pipelines in the open source space. Their SFT and RL setup is pulling ahead of 4.5-Air even approaching GLM-4.5 on CUDA coding tasks. The % difference is honestly impressive.


rasdani أعاد

Small PMPP-Eval update for freshly released Intellect-3 by @PrimeIntellect From my personal tests it was clear that its outperforming the Air variant (which uses same base model) numbers are confirming this with +%34 difference compared to Air and on par with 3x sized GLM-4.5

myainotez's tweet image. Small PMPP-Eval update for freshly released Intellect-3 by @PrimeIntellect 

From my personal tests it was clear that its outperforming the Air variant (which uses same base model) numbers are confirming this with +%34 difference compared to Air and on par with 3x sized GLM-4.5

United States الاتجاهات

Loading...

Something went wrong.


Something went wrong.