gitanol's profile picture. PhD in Computer Science

Willy Ugarte

@gitanol

PhD in Computer Science

Willy Ugarte memposting ulang

If you’re an AI engineer you should steal this pattern. Agent + State Management + Continuous Learning

⚡️⚡️Build a plan-and-learn agent with Agno and Gemini 3 Flash Most agents break when tasks get complex. They either overthink simple questions or lose track halfway through multi-step workflows. We got early access to Gemini 3 Flash and built something better.



Willy Ugarte memposting ulang

Yann LeCun (@ylecun ) beautifully explains how the architecture and principles used to train LLMs can not be extended to teach AI the real-world intelligence. In 1 line: LLMs excel where intelligence equals sequence prediction over symbols. Real-world intelligence requires…

Yann LeCun's new interview - explains why LLMs are so limited in terms of real-world intelligence. Says the biggest LLM is trained on about 30 trillion words, which is roughly 10 to the power 14 bytes of text. That sounds huge, but a 4 year old who has been awake about 16,000…



Willy Ugarte memposting ulang

⚔️ U-Net vs DiT For over 3 years, the good old U-Net was at the core of image generation. Early pixel-space diffusion, the Imagen series, Stable Diffusion 1.x, 2.x, and XL: all powered by this hybrid convolutional backbone. 1/N


Willy Ugarte memposting ulang

This paper from Harvard and MIT quietly answers the most important AI question nobody benchmarks properly: Can LLMs actually discover science, or are they just good at talking about it? The paper is called “Evaluating Large Language Models in Scientific Discovery”, and instead…

alex_prompter's tweet image. This paper from Harvard and MIT quietly answers the most important AI question nobody benchmarks properly:

Can LLMs actually discover science, or are they just good at talking about it?

The paper is called “Evaluating Large Language Models in Scientific Discovery”, and instead…

Willy Ugarte memposting ulang

This is insane 🤯 A new system called Paper2Video can read a scientific paper and automatically create a full presentation video slides, narration, subtitles, even a talking head of the author. It’s called PaperTalker, and it beat human-made videos in comprehension tests.…

ChrisLaubAI's tweet image. This is insane 🤯

A new system called Paper2Video can read a scientific paper and automatically create a full presentation video  slides, narration, subtitles, even a talking head of the author.

It’s called PaperTalker, and it beat human-made videos in comprehension tests.…

Willy Ugarte memposting ulang

Stanford just dropped their full LLM course on YouTube. 9 lectures. Completely Free. Real curriculum-level depth. CME 295: Transformers & Large Language Models This isn’t: • a hype tutorial • a prompt-engineering hack • a tech influencer hot take It’s Stanford’s Autumn…

srishticodes's tweet image. Stanford just dropped their full LLM course on YouTube.

9 lectures.
Completely Free.
Real curriculum-level depth.

CME 295: Transformers & Large Language Models

This isn’t: 
• a hype tutorial
• a prompt-engineering hack
• a tech influencer hot take

It’s Stanford’s Autumn…

Willy Ugarte memposting ulang

PhD Students - Here is an example of a good discussion section. A good discussion section should answer 6 questions. 1. What is different in your findings compared to previous research? 2. What is similar in your findings compared to previous research? 3. How different…

Faheem_uh's tweet image. PhD Students - Here is an example of a good discussion section.

A good discussion section should answer 6 questions.

1. What is different in your findings compared to previous research?

2. What is similar in your findings compared to previous research?

3. How different…

Willy Ugarte memposting ulang

so... free lunch is real? This paper shows that mathematically, euler discretization is removed (aka forgetting emerges naturally now), so this should remove any long context instability of linear attn? can anyone pls enlighten me if this is the case

bycloudai's tweet image. so... free lunch is real?

This paper shows that mathematically, euler discretization is removed (aka forgetting emerges naturally now), so this should remove any long context instability of linear attn?

can anyone pls enlighten me if this is the case
bycloudai's tweet image. so... free lunch is real?

This paper shows that mathematically, euler discretization is removed (aka forgetting emerges naturally now), so this should remove any long context instability of linear attn?

can anyone pls enlighten me if this is the case

With the recent hybrid attention releases from MiniMax, Qwen, Kimi, and NVIDIA, this paper introduces Error-Free Linear Attention that could top them all This new technique has a stable linear-time attention that's better than any linear attention variants and also DeltaNet!

askalphaxiv's tweet image. With the recent hybrid attention releases from MiniMax, Qwen, Kimi, and NVIDIA, this paper introduces Error-Free Linear Attention that could top them all

This new technique has a stable linear-time attention that's better than any linear attention variants and also DeltaNet!


Willy Ugarte memposting ulang

This guy literally shows how to master Claude Code from scratch (in 15 mins)


Willy Ugarte memposting ulang

Más del 90 % de los estudiantes ya usa IA de algún modo. Fingir que no existe no es una opción. El reto para escuelas y universidades es pasar del castigo al diseño: tareas auténticas, transparencia en el uso de IA y evaluación del razonamiento, no solo del producto final.…

sanz_ismael's tweet image. Más del 90 % de los estudiantes ya usa IA de algún modo. Fingir que no existe no es una opción. El reto para escuelas y universidades es pasar del castigo al diseño: tareas auténticas, transparencia en el uso de IA y evaluación del razonamiento, no solo del producto final.…

Willy Ugarte memposting ulang

This is the Outstanding Paper Award at ICLR 2025, and this is exactly the kind of research on LLMs we need, not those quasi-psychological studies of the form "we asked the same question to these 3 models and see which one is more racist!" As you might already know, when…

burkov's tweet image. This is the Outstanding Paper Award at ICLR 2025, and this is exactly the kind of research on LLMs we need, not those quasi-psychological studies of the form "we asked the same question to these 3 models and see which one is more racist!"

As you might already know, when…

Willy Ugarte memposting ulang

🚨BREAKING: You can now edit NotebookLM slides Codia just dropped NoteSlide, and it can convert NotebookLM slides to a fully editable PowerPoint deck in seconds. Free guide 👇

CodeByPoonam's tweet image. 🚨BREAKING: You can now edit NotebookLM slides

Codia just dropped NoteSlide, and it can convert NotebookLM slides to a fully editable PowerPoint deck in seconds.

Free guide 👇

Willy Ugarte memposting ulang

Multi-agent AI systems are eating single-agent architectures. But most teams have no idea how to build them. As agent systems get more complex, we're seeing a clear shift from single-agent architectures (one agent handles all tasks) to multi-agent architectures that distribute…

femke_plantinga's tweet image. Multi-agent AI systems are eating single-agent architectures.

But most teams have no idea how to build them.

As agent systems get more complex, we're seeing a clear shift from single-agent architectures (one agent handles all tasks) to multi-agent architectures that distribute…

Willy Ugarte memposting ulang

I just learned that Google solved HashMaps in C++ for everyone

ChShersh's tweet image. I just learned that Google solved HashMaps in C++ for everyone

Willy Ugarte memposting ulang

This guy literally shows how to go from zero to AI engineer for free

DBVolkov's tweet image. This guy literally shows how to go from zero to AI engineer for free

Willy Ugarte memposting ulang

This multi-agent system outperforms 9 of 10 human penetration testers. This work presents the first comprehensive evaluation of AI agents against human cybersecurity professionals on a real enterprise network: approximately 8,000 hosts across 12 subnets at a major research…

omarsar0's tweet image. This multi-agent system outperforms 9 of 10 human penetration testers.

This work presents the first comprehensive evaluation of AI agents against human cybersecurity professionals on a real enterprise network: approximately 8,000 hosts across 12 subnets at a major research…

Willy Ugarte memposting ulang

Stanford just made fine-tuning irrelevant with a single paper. It’s called Agentic Context Engineering (ACE) and it proves you can make models smarter without touching a single weight. Key takeaways (and get the 23 page PDF):

mdancho84's tweet image. Stanford just made fine-tuning irrelevant with a single paper.

It’s called Agentic Context Engineering (ACE) and it proves you can make models smarter without touching a single weight.

Key takeaways (and get the 23 page PDF):

Willy Ugarte memposting ulang

Builds AI agents locally without frameworks github.com/pguso/ai-agent…

tom_doerr's tweet image. Builds AI agents locally without frameworks

github.com/pguso/ai-agent…

Willy Ugarte memposting ulang

A free MIT course breaking down mathematics for computer science & engineering: bit.ly/3XPi0Ao (v/@MITOCW) Here, MIT prof. Tom Leighton, who is also the CEO & co-founder of Akamai, discusses what a proof is (Lesson 1).


Willy Ugarte memposting ulang

why does pre-norm work better than post-norm in transformers? i've been diving into transformer architecture (n'th time again) and noticed something interesting this time- almost all the implementations I have seen uses the "pre-norm" variant (normalizing before the sublayer…

viplismism's tweet image. why does pre-norm work better than post-norm in transformers?

i've been diving into transformer architecture (n'th time again) and noticed something interesting this time- almost all the implementations I have seen uses the "pre-norm" variant (normalizing before the sublayer…

Loading...

Something went wrong.


Something went wrong.