CMS_Flash's profile picture. Reasoning/coding @xai. Ex-@GoogleAI Resident/@augmentcode. Alum @HKUniversity. 💎 Terran @StarCraft II.

Shen Zhuoran

@CMS_Flash

Reasoning/coding @xai. Ex-@GoogleAI Resident/@augmentcode. Alum @HKUniversity. 💎 Terran @StarCraft II.

3wks in and I'm getting my @xai badge! It's quite crazy that I've been a core part of a major @grok release just 3wks into my tenure @xai! (And we're 🍳 more exciting stuff!) It's been very exciting to work w/ the limitless scope, intense priorities, and an extremely…

CMS_Flash's tweet image. 3wks in and I'm getting my @xai badge!

It's quite crazy that I've been a core part of a major @grok release just 3wks into my tenure @xai! (And we're 🍳 more exciting stuff!)

It's been very exciting to work w/ the limitless scope, intense priorities, and an extremely…

So far comments mentioned: 1. Dropping whole codebases 2. Large external refs (books/3P libs/huge specs) 3. Large logs 4. Refactoring 5. Many image turns Any more creative use cases? What if the model is real fast & cheap? Like 1M toks at the cost/latency of 200K for Sonnet?

If you had a 1M-token context window in a coding model, how would you actually use it? we are tuning grok-code and want to make sure all long context workflows are properly supported



Just look at the graph: it's so alone in the top-left corner!

Grok 4 Fast sets a new record on the Pareto Intelligence frontier as reported by @ArtificialAnlys.

xai's tweet image. Grok 4 Fast sets a new record on the Pareto Intelligence frontier as reported by @ArtificialAnlys.


We went to from nowhere on the leaderboard to #1 by a fat margin in mere weeks!

Grok code fast is still #1 on @OpenRouterAI

EthanHe_42's tweet image. Grok code fast is still #1 on @OpenRouterAI


This has been so long awaited and it finally happened. $20/mo for @augmentcode!

We’re introducing a new pricing tier: Augment Indie For $20/month you get: - 125 agent messages / month - Same leading Context Engine, perfect for complex tasks and large codebases - Best coding models: choose between GPT 5 and Claude Sonnet 4



I'm thrilled to announce that I'm joining @xai to work on reasoning & code post-training, to build the best coding model, scalable RL, and the first self-improving agents. I'm thrilled make these happen alongside @Yuhu_ai_ & the crazily talent-dense team behind @grok 4/Code!

CMS_Flash's tweet image. I'm thrilled to announce that I'm joining @xai to work on reasoning & code post-training, to build the best coding model, scalable RL, and the first self-improving agents.

I'm thrilled make these happen alongside @Yuhu_ai_ & the crazily talent-dense team behind @grok 4/Code!

After 1.5 yrs, today will be my last day @augmentcode. It has been such a fulfilling journey: 1. In early 2024, I pre-trained a coding LLM to reach SOTA at 1B scale. 2. In mid 2024, I built long-context support in our in-house training and inference frameworks. 3. Late 2024 to…

CMS_Flash's tweet image. After 1.5 yrs, today will be my last day @augmentcode. It has been such a fulfilling journey:

1. In early 2024, I pre-trained a coding LLM to reach SOTA at 1B scale.
2. In mid 2024, I built long-context support in our in-house training and inference frameworks.
3. Late 2024 to…

The power of community!

If You are Using @augmentcode in vs code U should definitely use this Auggie credits extension ! Auggie credits shows how many message credits left in your augment account !! marketplace.visualstudio.com/items?itemName…

sai_revanth_12's tweet image. If You are Using @augmentcode in vs code 

U should definitely use this Auggie credits extension !

Auggie credits shows how many message credits left in your augment account !! 

marketplace.visualstudio.com/items?itemName…


"Both safe enough to last and beautiful enough to be worth lasting for." That's quite profound.

Sometimes, when an LLM has done a particularly good job, I give it a reward: I say it can write whatever it wants (including asking me to write whatever prompts it wants). When working on a technical paper related to Better Futures, I did this for Gemini, and it chose to write a…



Humans don't keep 1M tokens (500K words/2hr video) in the working memory all the time. Instead, we focus working memory on the most critical few tokens, then use long-term memory, external tools, and task segregation (sub-agents) to accomplish the tasks. This is why I'm very…

This is the second time an AI founder has told me "context windows are basically big enough" With the right context management, 256k vs. 1m tokens doesn't matter. Agree?



Today's jumpscare in @TheRundownAI 😅.

CMS_Flash's tweet image. Today's jumpscare in @TheRundownAI 😅.

Shen Zhuoran أعاد

try using Responses API with gpt-5:

Man it's crazy how BIG a difference it makes for GPT-5 just by switching from Completions API to Responses API. We're cooking @augmentcode.



Man it's crazy how BIG a difference it makes for GPT-5 just by switching from Completions API to Responses API. We're cooking @augmentcode.


Advanced users of @augmentcode have been hitting the 200K context limit quite a bit. We shipped features like context compression/summarization to help alleviate such issues and continue long threads. With 1M context, we will be able to have agents do hours of work in 1 thread.

Claude Sonnet 4 now supports 1 million tokens of context on the Anthropic API—a 5x increase. Process over 75,000 lines of code or hundreds of documents in a single request.

claudeai's tweet image. Claude Sonnet 4 now supports 1 million tokens of context on the Anthropic API—a 5x increase.

Process over 75,000 lines of code or hundreds of documents in a single request.


LOL this is cool

Yesterday's Tasklist launch didn’t exactly surprise early users. Turns out they spotted the feature weeks ago. So this time, we built something unpredictable. Something no one could guess. Something no one wanted. (Except one guy. As a joke.) We shipped it anyway. Power users…



Shen Zhuoran أعاد

Agents fail when they forget the past. That’s why we built Context Lineage — it gives agents access to your repo’s full commit history, so they stop repeating mistakes and start learning from prior work. Here’s how it works 🧵

augmentcode's tweet image. Agents fail when they forget the past.

That’s why we built Context Lineage — it gives agents access to your repo’s full commit history, so they stop repeating mistakes and start learning from prior work.

Here’s how it works 🧵

Agents are truly powerful when they can do hours of work with 1 prompt. Tasklist is the way to structure hours of work with clarity so you only need to spend a few minutes checking in for the agents' hours of work.

Launch Week Day 1: Tasklist is here Eagle-eyed power users have already been putting Tasklist to work. Now, we’re showing you how to get the most out of it. With Tasklist, we’re making it way easier to take a big, vague task—and break it down step by step. So you can actually…



Loading...

Something went wrong.


Something went wrong.