Saurabh Bhatnagar
@analyticsaurabh
http://tora.bohita.com First Fashion Recommendation ML @ Rent The Runway 🦄, Founded ML at Barnes & Nobles. Past, @Virevol, Unilever, HP, ...
Bạn có thể thích
I just published “How I scaled Machine Learning to a Billion dollars: Strategy” medium.com/p/how-i-scaled…
In 2012 we took this wild ride at mobile infra at Facebook when trying to reduce the several-seconds long load time for “Newsfeed”. A few people worked on different approaches. Something we quickly realized was that setting up a connection with TCP and TLS was incredibly slow on…
Ghosts of my old startup for patent lawyers - tidylaw.com Also surprisingly, saving time is not a top law firm concern.
Agreed. Dwarkesh is just wrong here. GPT-5 Pro can now do legal research and analysis at a very high level (with limitations - may need to run even longer for certain searches; can't connect to proprietary databases). I use it to enhance my work all the time, with excellent…
Matches my experience; remote works better for senior engineers than junior ones (some caveats if they are unusually self-motivated)
Superstar labor econ team shows what we all should have expected: young workers learn more slowly when they work remote, experienced workers "can do their job" remote but partly by ignoring mentoring. Not in this empirical design, but I'd guess it applies to new *tasks* for all.
I say this with love, none of the modern paradigms were invented by HCI degree holders (sorry) They understand why something worked in the past better than anyone but… Much like historians
This feeling is what turned me off to doing academic datavis/HCI—obsessed with user studies, which always seemed very low yield and ceremonial. Like, we all have the same eyes, and I’ve thought about it more
My first job was in Ingres (non open source version) I dropped out of Columbia because it was more interesting than anything being taught. The planner/stats itself or balancing trees or real time check pointing are nerdswipes Later during Hadoop times (spark wasn’t out yet)…
Think databases are boring? Indexing → data structures + algorithms AI → RAG + learned indexes MVCC → concurrent programming Sharding → distributed systems Query parsing → formal languages Query planning → stats + optimization Replication → distributed systems…
When the ReAct paper came out I had a long argument with a Google FDE about why I think it’s just a bad flag plant because it doesn’t work But arguing against it would have required actually working through things instead of parroting what you read on socials
New on the Anthropic Engineering Blog: Long-running AI agents still face challenges working across many context windows. We looked to human engineers for inspiration in creating a more effective agent harness. anthropic.com/engineering/ef…
taste matters But to have taste you have to slog first
turns out, senior engineers accept more agent output than juniors. this is because: - they write higher-signal prompts with tighter spec and minimal ambiguity - they decompose work into agent-compatible units - they have stronger priors for correctness, making review faster and…
"One of the very confusing things about the models right now: how to reconcile the fact that they are doing so well on evals. And you look at the evals and you go, 'Those are pretty hard evals.' But the economic impact seems to be dramatically behind. There is [a possible]…
The @ilyasut episode 0:00:00 – Explaining model jaggedness 0:09:39 - Emotions and value functions 0:18:49 – What are we scaling? 0:25:13 – Why humans generalize better than models 0:35:45 – Straight-shotting superintelligence 0:46:47 – SSI’s model will learn from deployment…
in order to have research agents that can run for days, we need context compaction i used RL to have LLMs naturally learn their own 10x compression! Qwen learned to pack more info per token (ie use Mandarin tokens, prune text) read the technical blog: rajan.sh/llm-compression
nope! my weak-ish impression is that these take minimum of ~25 hours staff time, can be more like ~60 hours in rougher cases. some example time costs: - API issues - unclear ex ante which scaffold model will be most performant on in held-out dev task set - automatically +…
I don’t know about next 10 years But in 50 everyone will grow up with a private tutor ai and socialize for fun and trade ideas. This is how Newton, Darwin, Lovelace, Alexander, learnt in yore Schools are an artifact of needing daycare while in factories Colleges will thrive
A number of people are talking about implications of AI to schools. I spoke about some of my thoughts to a school board earlier, some highlights: 1. You will never be able to detect the use of AI in homework. Full stop. All "detectors" of AI imo don't really work, can be…
This is an unnerving statement and doesn’t match my experience with 4.1/3.5 And given that Adam is a programmer I respect, I have to consider the second order effects of me turning into a stone chiseler yelling at paper boys.
I believe this new model in Claude Code is a glimpse of the future we're hurtling towards, maybe as soon as the first half of next year: software engineering is done. Soon, we won't bother to check generated code, for the same reasons we don't check compiler output.
The way to AGI isn’t heavily RLed models on magic trick databases.
Nothing triggers me more when eval tools promote generic metrics (i.e. Affirmation, Brevity, Levenshtein) as way to make "evals easy" In reality, this is extremely poor data literacy sold as "best practices", in the same way that sugary cereal is marketed as healthy. The…
Microsoft launched plugins in Bing/Cooilot and featured us as one of 30 in the inaugural and only batch.
Signs of AGI
Huge support for adding litchi to unicode! It's a perfect example of digital discrimination / bias. Litchi is a very common and delicious fruit in many parts of the Asia yet none of the image generation model so far understands how to generate a peeled litchi. It seems that…
Excited to announce our MIT Press book “Neuroevolution: Harnessing Creativity in AI Agent Design” by Sebastian Risi (@risi1979), Yujin Tang (@yujin_tang), Risto Miikkulainen, and myself. We explore decades of work on evolving intelligent agents and shows how neuroevolution can…
On Nov 20, 2025 at 10:22 am EST Skynet went live
United States Xu hướng
- 1. Cowboys 73.2K posts
- 2. #heatedrivalry 23.7K posts
- 3. LeBron 104K posts
- 4. Gibbs 20K posts
- 5. Lions 90.7K posts
- 6. Pickens 14.4K posts
- 7. scott hunter 4,783 posts
- 8. #OnePride 10.5K posts
- 9. Brandon Aubrey 7,320 posts
- 10. fnaf 2 25.5K posts
- 11. Shang Tsung 27.8K posts
- 12. Ferguson 10.8K posts
- 13. Paramount 19.6K posts
- 14. Warner Bros 19.8K posts
- 15. CeeDee 10.5K posts
- 16. Eberflus 2,648 posts
- 17. Goff 8,682 posts
- 18. Bland 8,571 posts
- 19. #criticalrolespoilers 1,956 posts
- 20. #PowerForce N/A
Bạn có thể thích
-
Yao Fu
@Francis_YAO_ -
Jack Clark
@jackclarkSF -
Sylvain Gugger
@GuggerSylvain -
Gael Varoquaux 🦋
@GaelVaroquaux -
Sharon Zhou ✈️ NeurIPS
@realSharonZhou -
Dev Agrawal
@devagrawal09 -
Josh Gordon
@random_forests -
David Sussillo
@SussilloDavid -
Leandro von Werra
@lvwerra -
samim
@samim -
Dumitru Erhan
@doomie -
Leo Boytsov
@srchvrs -
Siyan Zhao @ NeurIPS
@siyan_zhao -
Ross Taylor
@rosstaylor90 -
Francesco Bottoni
@bot_fra
Something went wrong.
Something went wrong.