tech_optimist's profile picture. AI engineer @lancedb | prev @kuzudb. Blogging @ https://thedataquarry.com

Prashanth Rao

@tech_optimist

AI engineer @lancedb | prev @kuzudb. Blogging @ https://thedataquarry.com

Nice work! More and more folks are getting there, slowly, but surely 🚀


Somehow, I have a feeling that there's folks in my networks who are as excited about this as I am 😏 @joshmo_dev is inventing a new niche at the intersection of Rust + AI 🔥. Please grab his time!

next life goal: do a talk in the US about Rust and AI



Looking forward to doing my first talk (of many more to come) about @lancedb and also to meeting all the coolest folks working on retrieval, agents and RAG! 👇🏽 Will be doing a *lot* more thinking after such an event, I'm sure 😅

Prashanth Rao @tech_optimist, AI Engineer, is speaking at @ScaleByTheBay next week! Don’t miss his session - Agents Need More Than Vector Search - on Tuesday 11/18. This talk will introduce the principles behind LanceDB and show how its AI-native multimodal lakehouse can power…

lancedb's tweet image. Prashanth Rao @tech_optimist, AI Engineer, is speaking at @ScaleByTheBay next week!

Don’t miss his session - Agents Need More Than Vector Search - on Tuesday 11/18.

This talk will introduce the principles behind LanceDB and show how its AI-native multimodal  lakehouse can power…


Is big data getting TOO big? What's up with @github going down every other week?


Awesome recent demo by Pablo Delgado from @Netflix! Searching over video has never been easier (or more scalable) thanks to @lancedb. The demo shows how to use natural language queries to retrieve relevant video frames from petabytes of video, with <1s latency. 🤯 👇🏽👇🏽👇🏽

Here's a demo from @Netflix's Pablo Delgado @pablete showcasing text-to-text semantic search at Ray Summit 2025. Watch how natural language queries like "elephant in the desert" instantly retrieve relevant video frames from petabytes of content—with sub-second latency. 1/3



Prashanth Rao reposted

when there's a shovel rush, dig for gold


Here's a gem for the job market: I for one would love to see Herumb @krypticmouse at places like @modal or @cerebras or any of the other cool places that are building the next generation of AI & ML training/inference infra. When Herumb puts his systems programming hat on, he's a…

I just learned that the one-of-a-kind Herumb Shandilya @krypticmouse, who is soon to graduate from Stanford with Master’s in CS, is on the job market for training infra, inference infra, or other ML engineering roles. Here’s a PSA to y’all: you should hire him.



There are so many interesting research directions that are already being taken in this front, and I'm eager to see how it all pans out over the coming years: - Separating a model's reasoning from its memory (they're currently intertwined) - Selectively "forgetting" from a model's…

What's the latest research on how much baked-in knowledge an LLM needs in order to be useful? If I want a specialist coding model can I trim the size of that model down by stripping out detailed knowledge of human history, geography etc? Can we even do that?



Prashanth Rao reposted

A really great long-context benchmark!

Can LLMs accurately aggregate information over long, information-dense texts? Not yet… We introduce Oolong, a dataset of simple-to-verify information aggregation questions over long inputs. No model achieves >50% accuracy at 128K on Oolong!

abertsch72's tweet image. Can LLMs accurately aggregate information over long, information-dense texts? Not yet…

We introduce Oolong, a dataset of simple-to-verify information aggregation questions over long inputs. No model achieves &amp;gt;50% accuracy at 128K on Oolong!


When it comes to open source, locally-stored, private memories for your application, @lancedb is great option. One Q I have is this: how is it decided at the application layer when something should be "forgotten"? Worth checking out AVDB by @Shashikant86!👇🏽

Still not sure why big AI labs still rely on the File System, Virtual File System for Agent Memory (e.g Claude Code), Whats issue with using Vector Store like @lancedb as agent memory? It's impressive DiskANN can be game changing for context/memory management. The first project I…

Shashikant86's tweet image. Still not sure why big AI labs still rely on the File System, Virtual File System for Agent Memory (e.g Claude Code), Whats issue with using Vector Store like @lancedb as agent memory? It&apos;s impressive DiskANN can be game changing for context/memory management. The first project I…


The number of times I've felt this in my lifetime is non-trivial

why are you, someone who doesnt hate themselves, using databases written in java in 2025?



I'm unable to predict what will run out sooner: entertaining posts about embedding-based retrieval on HN, or ideas (and the money involved to fund them) in Silicon Valley

So much confusion around embedding-based retriveal and associated technologies that it will continue to provide entertainment on hackernews for decades



Loading...

Something went wrong.


Something went wrong.