mikecarroll_eng's profile picture. Engineer. Previously @Facebook

Mike Carroll

@mikecarroll_eng

Engineer. Previously @Facebook

🤣

new sitcom idea: 18 yo chinese ccp spy joining an SF defense startup and realizing no one has actually done any work because they are too busy going to parties and chasing social clout, so she has to create all of the tech herself before she can send it back home



Mike Carroll أعاد

>Augmented reality + generative AGI means people can suddenly do expert-level work they never trained it's already here

Two underdiscussed possibilities: Augmented reality + generative AGI means people can suddenly do expert-level work they never trained for. Real-time overlays guide you through any task - repairs, construction, technical assembly, complex procedures. Visual guides showing…



Mike Carroll أعاد

Starship vs A380 & 737-8 When stacked with the Super Heavy booster Starship is taller than an A380 and a Max-8 on top of each other, 123m vs 73m+40m

ApoStructura's tweet image. Starship vs A380 & 737-8

When stacked with the Super Heavy booster Starship is taller than an A380 and a Max-8 on top of each other, 123m vs 73m+40m

Starship vs 737-8 Easy to forget how massive Starship is, 52m long vs 40m for a Max-8

ApoStructura's tweet image. Starship vs 737-8

Easy to forget how massive Starship is, 52m long vs 40m for a Max-8


Mike Carroll أعاد

18 months ago, @karpathy set a challenge: "Can you take my 2h13m tokenizer video and translate [into] a book chapter". We've done it! It includes prose, code & key images. It's a great way to learn this key piece of how LLMs work. fast.ai/posts/2025-10-…


Mike Carroll أعاد

Video lectures, MIT 6.824 Distributed Systems spring 2020, by Robert Morris nil.csail.mit.edu/6.824/2020/gen… youtube.com/playlist?list=…


Mike Carroll أعاد

ML concepts every data scientist should know for interviews: Bookmark this. 1. Bias-Variance Tradeoff 2. Cross-Validation Strategies 3. Regularization (L1, L2, Elastic Net) 4. Class Imbalance & Sampling Techniques 5. Feature Engineering & Selection 6. Overfitting vs…


Mike Carroll أعاد

Training Andrej Karpathy’s Nanochat on 4x RTX 3090s at 225W each: Step 2,694/21,400 (12.59% done) Loss: 3.14 Runtime: 6.78 hours Throughput: 3,600 tok/sec Temps: 52-57°C VRAM: 19GB/24GB per card Total cost: 15$ at 55h Zero errors, perfectly stable

0x_Sero's tweet image. Training Andrej Karpathy’s Nanochat on 4x RTX 3090s at 225W each:

Step 2,694/21,400 (12.59% done)
Loss: 3.14
Runtime: 6.78 hours
Throughput: 3,600 tok/sec
Temps: 52-57°C
VRAM: 19GB/24GB per card
Total cost: 15$ at 55h

Zero errors, perfectly stable
0x_Sero's tweet image. Training Andrej Karpathy’s Nanochat on 4x RTX 3090s at 225W each:

Step 2,694/21,400 (12.59% done)
Loss: 3.14
Runtime: 6.78 hours
Throughput: 3,600 tok/sec
Temps: 52-57°C
VRAM: 19GB/24GB per card
Total cost: 15$ at 55h

Zero errors, perfectly stable

Mike Carroll أعاد

I don't like courses. Most were a waste of time. Yes, even at Stanford. If you're new to ML, take CS231N.

your honor i object, i dont know about harvard but stanford literally releases SOTA courses

dejavucoder's tweet image. your honor i object, i dont know about harvard but stanford literally releases SOTA courses
dejavucoder's tweet image. your honor i object, i dont know about harvard but stanford literally releases SOTA courses


Mike Carroll أعاد

My meeting budget: 5 min - meet someone new 10 min - solve a problem 15 min - identify + solve a problem Parkinson’s law: work expands so as to fill the time available for its completion.


Mike Carroll أعاد

MIT's 6.851: Advanced Data Structures (Spring'21) courses.csail.mit.edu/6.851/spring21/ This has been on my recommendation list for a while, and the Memory hierarchy discussions are great in the context of cache-oblivious algorithms.

vivekgalatage's tweet image. MIT's 6.851: Advanced Data Structures (Spring'21)

courses.csail.mit.edu/6.851/spring21/

This has been on my recommendation list for a while, and the Memory hierarchy discussions are great in the context of cache-oblivious algorithms.

"Cache‑Oblivious Algorithms and Data Structures" by Erik D. Demaine erikdemaine.org/papers/BRICS20… This is a foundational survey on designing cache‑oblivious algorithms and data structures that perform as well as cache‑aware approaches that require hardcoding cache size (M) and block…

vivekgalatage's tweet image. "Cache‑Oblivious Algorithms and Data Structures" by Erik D. Demaine

erikdemaine.org/papers/BRICS20…

This is a foundational survey on designing cache‑oblivious algorithms and data structures that perform as well as cache‑aware approaches that require hardcoding cache size (M) and block…


Mike Carroll أعاد

50 LLM Projects with Source Code to Become a Pro 1. Beginner-Level LLM Projects → Text Summarizer using OpenAI API → Chatbot for Customer Support → Sentiment Analysis with GPT Models → Resume Optimizer using LLMs → Product Description Generator → AI-Powered Grammar…

e_opore's tweet image. 50 LLM Projects with Source Code to Become a Pro

1. Beginner-Level LLM Projects

→ Text Summarizer using OpenAI API
→ Chatbot for Customer Support
→ Sentiment Analysis with GPT Models
→ Resume Optimizer using LLMs
→ Product Description Generator
→ AI-Powered Grammar…

Mike Carroll أعاد

I left my plans for weekend to read this recent blog from HuggingFace 🤗 on how they maintain the most critical AI library: transformers. → 1M lines of Python, → 1.3M installations, → thousands of contributors, → a true engineering masterpiece, Here's what I learned:…

Hesamation's tweet image. I left my plans for weekend to read this recent blog from HuggingFace 🤗 on how they maintain the most critical AI library: transformers.

→ 1M lines of Python,
→ 1.3M installations,
→ thousands of contributors,
→ a true engineering masterpiece, 

Here's what I learned:…

Mike Carroll أعاد

Excited to release new repo: nanochat! (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…

karpathy's tweet image. Excited to release new repo: nanochat!
(it's among the most unhinged I've written).

Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…

Mike Carroll أعاد

You're not depressed, you just lost your quest.

NTFabiano's tweet image. You're not depressed, you just lost your quest.

Mike Carroll أعاد

🔥Free Google Collab notebooks to implement every Machine Learning Algorithm from scratch Link in comment

victor_explore's tweet image. 🔥Free Google Collab notebooks to implement every Machine Learning Algorithm from scratch Link in comment

Mike Carroll أعاد

how i got here: > i used to be and still tend towards having an obsessive/addictive personality > put many years of my life into video games > it was only 2 years ago i started to turn that around because i got other interests and starting really looking forward to the future >…


Mike Carroll أعاد

found a repo that has a massive collection of Machine Learning system design case studies used in the real world, from Stripe, Spotify, Netflix, Meta, GitHub, Twitter/X, and much more link in replies

d4rsh_tw's tweet image. found a repo that has a massive collection of Machine Learning system design case studies used in the real world, from Stripe, Spotify, Netflix, Meta, GitHub, Twitter/X, and much more

link in replies

Mike Carroll أعاد

Copy-pasting PyTorch code is fast — using an AI coding model is even faster — but both skip the learning. That's why I asked my students to write by hand ✍️. 🔽 Download: byhand.ai/pytorch After the exercise, my students can understand what every line really does and…


Mike Carroll أعاد

70 Python Projects with Source code for Developers Step 1: Beginner Foundations → Hello World Web App → Calculator (CLI) → To-Do List CLI → Number Guessing Game → Countdown Timer → Dice Roll Simulator → Coin Flip Simulator → Password Generator → Palindrome Checker →…


Loading...

Something went wrong.


Something went wrong.