ManoukDrai47041's profile picture. Agent Simulations are the new unit testing.

Manouk Draisma

@ManoukDrai47041

Agent Simulations are the new unit testing.

Manouk Draisma reposted

In agent demos, everything’s smooth. In prod? You get messy inputs, long chains, weird edge cases — that’s when things snap. We treat agents like code → write scenario tests first, simulate full workflows, then iterate until green. Think TDD, but for LLMs. More on how we do it…


Manouk Draisma reposted

“Do I really need evals?” The real q: how do you know your AI agents will behave in prod? Prototypes don’t need them. Scaling products do. That’s why we built Agent Simulations; Unit tests for AI. The only way to know if you can ship reliably. OSS: github.com/langwatch/scen…


Manouk Draisma reposted

AI SDK Observability Integration: @LangWatchAI Observability and evals are crucial when developing AI applications. You can use LangWatch with the AI SDK to monitor and evaluate your LLM calls:

aisdk's tweet image. AI SDK Observability Integration: @LangWatchAI

Observability and evals are crucial when developing AI applications. You can use LangWatch with the AI SDK to monitor and evaluate your LLM calls:

United States Trends

Loading...

Something went wrong.


Something went wrong.