code_igx's profile picture. 25 🇮🇳, Hustler @RITtigers NY 🇺🇸 | RnD on Quantum AI, Superintelligence & Systems | Ex- @Broadcom @VMware

Ishan Gupta

@code_igx

25 🇮🇳, Hustler @RITtigers NY 🇺🇸 | RnD on Quantum AI, Superintelligence & Systems | Ex- @Broadcom @VMware

Ishan Gupta podał dalej

🚨 This might be the biggest leap in AI agents since ReAct. Researchers just dropped DeepAgent a reasoning model that can think, discover tools, and act completely on its own. No pre-scripted workflows. No fixed tool lists. Just pure autonomous reasoning. It introduces…

rryssf_'s tweet image. 🚨 This might be the biggest leap in AI agents since ReAct.

Researchers just dropped DeepAgent a reasoning model that can think, discover tools, and act completely on its own.

No pre-scripted workflows. No fixed tool lists. Just pure autonomous reasoning.

It introduces…

Ishan Gupta podał dalej

7 patterns to build multi-agent systems:

_avichawla's tweet image. 7 patterns to build multi-agent systems:

Ishan Gupta podał dalej

𝗟𝗘𝗔𝗞𝗘𝗗: 100s of premium AI Agents... These exact Agents sell for $𝟱,𝟬𝟬𝟬+ 𝗽𝗲𝗿 𝗯𝘂𝗶𝗹𝗱, 𝗲𝗮𝘀𝗶𝗹𝘆... Inside the file you get: → Lead qualification agents → Content generation pipelines → Appointment booking automation → Cold outreach sequences → Data…


Ishan Gupta podał dalej

Supervised Reinforcement Learning: From Expert Trajectories to Step-wise Reasoning Breaks down SFT dataset demonstrations into a sequence of actions, generate internal reasoning before each action, reward based on similarity of model's actions and expert actions. Experiments…

iScienceLuvr's tweet image. Supervised Reinforcement Learning: From Expert Trajectories to Step-wise Reasoning

Breaks down SFT dataset demonstrations into a sequence of actions, generate internal reasoning before each action, reward based on similarity of model's actions and expert actions. Experiments…

Ishan Gupta podał dalej

The paper shows how Human-Computer Interaction (HCI) talks about LLM reasoning without looking at what builds it. The authors read 258 CHI papers from 2020 to 2025 to see how reasoning is used . They find many papers use reasoning as a selling point to try an idea with LLMs.…

rohanpaul_ai's tweet image. The paper shows how Human-Computer Interaction (HCI)  talks about LLM reasoning without looking at what builds it.

The authors read 258 CHI papers from 2020 to 2025 to see how reasoning is used .

They find many papers use reasoning as a selling point to try an idea with LLMs.…

Ishan Gupta podał dalej

Scaling Latent Reasoning via Looped Language Models 1.4B and 2.6B param LoopLMs pretrained on 7.7T tokens match the performance of 4B and 8B standard transformers respectively across nearly all benchmarks time to be bullish on adaptive computation again? great work by…

iScienceLuvr's tweet image. Scaling Latent Reasoning via Looped Language Models

1.4B and 2.6B param LoopLMs pretrained on 7.7T tokens match the performance of 4B and 8B standard transformers respectively across nearly all benchmarks

time to be bullish on adaptive computation again?

great work by…

Ishan Gupta podał dalej

Graph-based Agent Planning It lets AI agents run multiple tools in parallel to accelerate task completion. Uses graphs to map tool dependencies + RL to learn the best execution order. RL also helps with scheduling strategies and planning. Major speedup for complex tasks.

omarsar0's tweet image. Graph-based Agent Planning

It lets AI agents run multiple tools in parallel to accelerate task completion.

Uses graphs to map tool dependencies + RL to learn the best execution order.

RL also helps with scheduling strategies and planning.

Major speedup for complex tasks.

Ishan Gupta podał dalej

New Nvidia paper shows how a single LLM can teach itself to reason better. It creates 3 roles from the same model, a Proposer, a Solver, and a Judge. The Proposer writes hard but solvable questions that stretch the model. The Solver answers those questions with clear steps and…

rohanpaul_ai's tweet image. New Nvidia paper shows how a single LLM can teach itself to reason better.

It creates 3 roles from the same model, a Proposer, a Solver, and a Judge.

The Proposer writes hard but solvable questions that stretch the model.

The Solver answers those questions with clear steps and…

Ishan Gupta podał dalej

Tesla autonomous driving might spread faster than any technology ever. The hardware foundations have been laid for such a long time that a software update enables self-driving for millions of pre-existing cars in a short period of time.

Comparison of Tesla's vs Waymo's Robotaxi geofence map in Austin, Texas. Today, @Tesla expanded their geofence area for the first time in two months.

SawyerMerritt's tweet image. Comparison of Tesla's vs Waymo's Robotaxi geofence map in Austin, Texas.

Today, @Tesla expanded their geofence area for the first time in two months.


Ishan Gupta podał dalej

MCP & A2A (Agent2Agent) protocol, clearly explained! Agentic applications require both A2A and MCP. - MCP provides agents with access to tools. - A2A allows agents to connect with other agents and collaborate in teams. Let's understand what A2A is and how it can work with MCP:…

_avichawla's tweet image. MCP & A2A (Agent2Agent) protocol, clearly explained!

Agentic applications require both A2A and MCP.

- MCP provides agents with access to tools.
- A2A allows agents to connect with other agents and collaborate in teams.

Let's understand what A2A is and how it can work with MCP:…

Ishan Gupta podał dalej

holy sh*t... your llm remembers everything you typed 🤯 researchers just proved you can recover the EXACT input text from a language model's hidden states. not similar text. not approximate. the actual words you typed. here's what they found: • transformer language models…

alex_prompter's tweet image. holy sh*t... your llm remembers everything you typed 🤯

researchers just proved you can recover the EXACT input text from a language model's hidden states. 

not similar text. not approximate. 

the actual words you typed.

here's what they found:
• transformer language models…

Ishan Gupta podał dalej

The goal of Grok and Grokipedia.com is the truth, the whole truth and nothing but the truth. We will never be perfect, but we shall nonetheless strive towards that goal.

Note the difference between Wikipedia's first paragraph on George Floyd compared to the first paragraph from Grokipedia. The nuance and detail on Grokipedia is FAR superior to Wikipedia and is clearly not pushing any ideologies, unlike Wikipedia. Corrections like this are…

DillonLoomis's tweet image. Note the difference between Wikipedia's first paragraph on George Floyd compared to the first paragraph from Grokipedia. The nuance and detail on Grokipedia is FAR superior to Wikipedia and is clearly not pushing any ideologies, unlike Wikipedia. Corrections like this are…
DillonLoomis's tweet image. Note the difference between Wikipedia's first paragraph on George Floyd compared to the first paragraph from Grokipedia. The nuance and detail on Grokipedia is FAR superior to Wikipedia and is clearly not pushing any ideologies, unlike Wikipedia. Corrections like this are…


Ishan Gupta podał dalej

SpaceX is now launching roughly the same number of satellites in a year as the number of operational non-SpaceX satellites in orbit cumulatively

SpaceX has officially surpassed its total number of launches from all of 2024 (138) — with more than two months still left in 2025. The company has launched over 2,500 Starlink satellites in 2025 alone. An absolutely incredible pace of progress 🚀

SawyerMerritt's tweet image. SpaceX has officially surpassed its total number of launches from all of 2024 (138) — with more than two months still left in 2025. The company has launched over 2,500 Starlink satellites in 2025 alone.

An absolutely incredible pace of progress 🚀
SawyerMerritt's tweet image. SpaceX has officially surpassed its total number of launches from all of 2024 (138) — with more than two months still left in 2025. The company has launched over 2,500 Starlink satellites in 2025 alone.

An absolutely incredible pace of progress 🚀
SawyerMerritt's tweet image. SpaceX has officially surpassed its total number of launches from all of 2024 (138) — with more than two months still left in 2025. The company has launched over 2,500 Starlink satellites in 2025 alone.

An absolutely incredible pace of progress 🚀
SawyerMerritt's tweet image. SpaceX has officially surpassed its total number of launches from all of 2024 (138) — with more than two months still left in 2025. The company has launched over 2,500 Starlink satellites in 2025 alone.

An absolutely incredible pace of progress 🚀


Ishan Gupta podał dalej

I got rejected by 144 investors before raising $150M for my $200M+ rev/year startup. After 144 rejections, I started questioning our approach. Were we solving the right problem? What were we doing wrong? Why weren’t investors seeing what we were seeing? Were we the right…


Ishan Gupta podał dalej

This is actually a clever context engineering technique for web agents. It's called AgentFold, an agent that acts as a self-aware knowledge manager. It treats context as a dynamic cognitive workspace by folding information at different scales: - Light folding: Compressing…

omarsar0's tweet image. This is actually a clever context engineering technique for web agents.

It's called AgentFold, an agent that acts as a self-aware knowledge manager.

It treats context as a dynamic cognitive workspace by folding information at different scales:

- Light folding: Compressing…
omarsar0's tweet image. This is actually a clever context engineering technique for web agents.

It's called AgentFold, an agent that acts as a self-aware knowledge manager.

It treats context as a dynamic cognitive workspace by folding information at different scales:

- Light folding: Compressing…
omarsar0's tweet image. This is actually a clever context engineering technique for web agents.

It's called AgentFold, an agent that acts as a self-aware knowledge manager.

It treats context as a dynamic cognitive workspace by folding information at different scales:

- Light folding: Compressing…

Ishan Gupta podał dalej

wrote a blog on LLM quantization, link in comments

prathamgrv's tweet image. wrote a blog on LLM quantization, link in comments

Ishan Gupta podał dalej

Introducing Multi-Agent Evolve 🧠 A new paradigm beyond RLHF and RLVR: More compute → closer to AGI No need for expensive data or handcrafted rewards We show that an LLM can self-evolve — improving itself through co-evolution among roles (Proposer, Solver, Judge) via RL — all…

youjiaxuan's tweet image. Introducing Multi-Agent Evolve 🧠

A new paradigm beyond RLHF and RLVR:
More compute → closer to AGI
No need for expensive data or handcrafted rewards

We show that an LLM can self-evolve — improving itself through co-evolution among roles (Proposer, Solver, Judge) via RL — all…
youjiaxuan's tweet image. Introducing Multi-Agent Evolve 🧠

A new paradigm beyond RLHF and RLVR:
More compute → closer to AGI
No need for expensive data or handcrafted rewards

We show that an LLM can self-evolve — improving itself through co-evolution among roles (Proposer, Solver, Judge) via RL — all…
youjiaxuan's tweet image. Introducing Multi-Agent Evolve 🧠

A new paradigm beyond RLHF and RLVR:
More compute → closer to AGI
No need for expensive data or handcrafted rewards

We show that an LLM can self-evolve — improving itself through co-evolution among roles (Proposer, Solver, Judge) via RL — all…

Ishan Gupta podał dalej

🧵 LoRA vs full fine-tuning: same performance ≠ same solution. Our NeurIPS ‘25 paper 🎉shows that LoRA and full fine-tuning, even when equally well fit, learn structurally different solutions and that LoRA forgets less and can be made even better (lesser forgetting) by a simple…

ReeceShuttle's tweet image. 🧵 LoRA vs full fine-tuning: same performance ≠ same solution.

Our NeurIPS ‘25 paper 🎉shows that LoRA and full fine-tuning, even when equally well fit, learn structurally different solutions and that LoRA forgets less and can be made even better (lesser forgetting) by a simple…

Ishan Gupta podał dalej

Just saw the MiniMax-M2 benchmarks, and the performance is too good to ignore :). So, I just amended my "The Big LLM Architecture Comparison" with entry number 13! 1️⃣ Full attention modules: As shown in the overview figure below, I grouped MiniMax-M2 with the other…

rasbt's tweet image. Just saw the MiniMax-M2 benchmarks, and the performance is too good to ignore :). So, I just amended my "The Big LLM Architecture Comparison" with entry number 13! 

1️⃣ Full attention modules:

As shown in the overview figure below, I grouped MiniMax-M2 with the other…

Ishan Gupta podał dalej

I finally understand the fundamentals of building real AI agents. This new paper “Fundamentals of Building Autonomous LLM Agents” breaks it down so clearly it feels like a blueprint for digital minds. Turns out, true autonomy isn’t about bigger models. It’s about giving an LLM…

AlexanderFYoung's tweet image. I finally understand the fundamentals of building real AI agents.

This new paper “Fundamentals of Building Autonomous LLM Agents” breaks it down so clearly it feels like a blueprint for digital minds.

Turns out, true autonomy isn’t about bigger models.

It’s about giving an LLM…

Loading...

Something went wrong.


Something went wrong.