_mihirpatel67's profile picture. || ॐ || Data Geek || #AI #ML 👨🏻‍💻 Data Scientist ||    I 😍 Reading

M

@_mihirpatel67

|| ॐ || Data Geek || #AI #ML 👨🏻‍💻 Data Scientist || I 😍 Reading

Pinned

¯\_(ツ)_/¯ idk. idc. idgaf. ¯\_(ツ)_/¯


M reposted

Come check out Agent Laboratory at #EMNLP2025! We will discuss how LLM Agents can help accelerate scientific discovery. I'll be presenting on Nov 5 @ 8:00. Looking forward to the discussions!

SRSchmidgall's tweet image. Come check out Agent Laboratory at #EMNLP2025!

We will discuss how LLM Agents can help accelerate scientific discovery. I'll be presenting on Nov 5 @ 8:00.

Looking forward to the discussions!

M reposted

The Illustrated NeurIPS 2025: A Visual Map of the AI Frontier New blog post! NeurIPS 2025 papers are out—and it’s a lot to take in. This visualization lets you explore the entire research landscape interactively, with clusters, summaries, and @cohere LLM-generated explanations…


M reposted

MIT published a 26-page report on AI. Save this one.

aaditsh's tweet image. MIT published a 26-page report on AI. Save this one.
aaditsh's tweet image. MIT published a 26-page report on AI. Save this one.

M reposted

This thought is converging from many sides. Transformer based LLMs are not going take us to human level AI. That famous Yann LeCun interview. "We are not going to get to human level AI by just scaling up MLMs. This is just not going to happen. There's no way. Okay, absolutely…

Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you.. There is a 3D world that follows laws of physics." Language is purely generated signal.



M reposted

Derivation of time-dependent Schrödinger's equation using wave mechanics ✍️

PhysInHistory's tweet image. Derivation of time-dependent Schrödinger's equation using wave mechanics ✍️

M reposted

💥 Top 50 LLM Interview Questions & Answers Get ready for your next AI / ML / LLM interview with this power-packed Q&A guide covering: ✅ Prompt Engineering ✅ Fine-tuning & RAG ✅ Transformer Architecture ✅ Tokenization & Attention ✅ Real-world LLM Scenarios Perfect for…

Krishnasagrawal's tweet image. 💥 Top 50 LLM Interview Questions & Answers

Get ready for your next AI / ML / LLM interview with this power-packed Q&A guide covering:
✅ Prompt Engineering
✅ Fine-tuning & RAG
✅ Transformer Architecture
✅ Tokenization & Attention
✅ Real-world LLM Scenarios

Perfect for…
Krishnasagrawal's tweet image. 💥 Top 50 LLM Interview Questions & Answers

Get ready for your next AI / ML / LLM interview with this power-packed Q&A guide covering:
✅ Prompt Engineering
✅ Fine-tuning & RAG
✅ Transformer Architecture
✅ Tokenization & Attention
✅ Real-world LLM Scenarios

Perfect for…

M reposted

Highly recommend this work from @shannonzshen if you're interested in human–agent collaboration but have felt discouraged by how hard it is to quantitatively study collaboration with humans in the loop. To quantify human-agent collaboration, the work introduces…

EchoShao8899's tweet image. Highly recommend this work from @shannonzshen if you're interested in human–agent collaboration but have felt discouraged by how hard it is to quantitatively study collaboration with humans in the loop.

To quantify human-agent collaboration, the work introduces…

Today's AI agents are optimized to complete tasks in one shot. But real-world tasks are iterative, with evolving goals that need collaboration with users. We introduce collaborative effort scaling to evaluate how well agents work with people—not just complete tasks 🧵

shannonzshen's tweet image. Today's AI agents are optimized to complete tasks in one shot. But real-world tasks are iterative, with evolving goals that need collaboration with users.

We introduce collaborative effort scaling to evaluate how well agents work with people—not just complete tasks 🧵


M reposted

If you want to increase your surface area for luck, focus on producing proof of work. You will make yourself a bigger target for luck.

Kpaxs's tweet image. If you want to increase your surface area for luck, focus on producing proof of work. You will make yourself a bigger target for luck.

M reposted

Stop Prompting LLMs. Start Programming LLMs. Introducing DSPy by Stanford NLP. This is why you need to learn it:

mdancho84's tweet image. Stop Prompting LLMs. 
Start Programming LLMs.

Introducing DSPy by Stanford NLP. 

This is why you need to learn it:

M reposted

step-by-step LLM Engineering Projects each project = one concept learned the hard (i.e. real) way Tokenization & Embeddings > build byte-pair encoder + train your own subword vocab > write a “token visualizer” to map words/chunks to IDs > one-hot vs learned-embedding: plot…


M reposted

This broke my brain. A team at Sea AI Lab just discovered that most of the chaos in reinforcement learning training collapse, unstable gradients, inference drift wasn’t caused by the algorithms at all. It was caused by numerical precision. The default BF16 format, used across…

ChrisLaubAI's tweet image. This broke my brain.

A team at Sea AI Lab just discovered that most of the chaos in reinforcement learning  training collapse, unstable gradients, inference drift wasn’t caused by the algorithms at all.

It was caused by numerical precision.

The default BF16 format, used across…

M reposted

This guy literally teaches you how to learn complex skills

aibytekat's tweet image. This guy literally teaches you how to learn complex skills

M reposted

Simplicity is at the heart of great software. This is one of the reasons why Claude Code has been sticky for me. As a builder, I love planning and brainstorming, and this is now a key focus of Claude Code. I use Shift + Tab a lot to cycle between brainstorming, planning, and…


M reposted

Elon’s famous analogy: “Every person in your company is a vector. Your progress is determined by the sum of all vectors”. At big companies, vectors often point to different directions. The sum is zero. xAI has a small team of engineers, but every vector points at the same…


M reposted

bf16 doesn’t work when you do sft on qwen-3-1.7b on a multi-turn wordle env. reward stays at 0.2 whereas with fp32 it goes to ~0.65.

eigenron's tweet image. bf16 doesn’t work when you do sft on qwen-3-1.7b on a multi-turn wordle env. reward stays at 0.2 whereas with fp32 it goes to ~0.65.

M reposted

ai engineer interview question

lochan_twt's tweet image. ai engineer interview question

ai internship DSA question

lochan_twt's tweet image. ai internship DSA question


M reposted

Fei-Fei Li says language models are extremely limited. This @GoogleDeepMind paper makes almost the same point, just in the world of video. The models are just very advanced pattern matchers. They can recreate what looks like reality because they’ve seen so much data, but they…

rohanpaul_ai's tweet image. Fei-Fei Li says language models are extremely limited.

This @GoogleDeepMind paper makes almost the same point, just in the world of video.

The models are just very advanced pattern matchers. They can recreate what looks like reality because they’ve seen so much data, but they…

Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you.. There is a 3D world that follows laws of physics." Language is purely generated signal.



M reposted

"Mathematical Foundations of Machine Learning" PDF: nowak.ece.wisc.edu/MFML.pdf

Riazi_Cafe_en's tweet image. "Mathematical Foundations of Machine Learning"  

PDF: nowak.ece.wisc.edu/MFML.pdf

M reposted

i used the following resources to understand the quantization process for NVFP4, the paper below made it very clear. - cublas docs: docs.nvidia.com/cuda/cublas/in… - arxiv paper (appendix B): arxiv.org/pdf/2509.25149 - OCP microscaling spec: opencompute.org/documents/ocp-…

simple NVFP4 quantization within 100 lines of pytorch

mrsiipa's tweet image. simple NVFP4 quantization within 100 lines of pytorch


M reposted

LoRA by Hand ✍️ ~ How does LoRA make fine-turning efficient (fewer parameters)? I drew this animation to show you. Pay attention to the number of red cells. On the left, regular fine-tune would involve 10x10=100 trainable parameters to update the original weight matrix.. On…


United States Trends

Loading...

Something went wrong.


Something went wrong.