code_star's profile picture. Data Dawg @datologyai | Formerly Data Research Lead @DbrxMosaicAI | Visiting Researcher @ Facebook | Ph.D | #TXSTFOOTBALL fan | http://linktr.ee/code_star

Cody Blakeney

@code_star

Data Dawg @datologyai | Formerly Data Research Lead @DbrxMosaicAI | Visiting Researcher @ Facebook | Ph.D | #TXSTFOOTBALL fan | http://linktr.ee/code_star

고정된 트윗

We are looking for a post-training lead at @datologyai we have gpus, you can make them go brrrr

code_star's tweet image. We are looking for a post-training lead at @datologyai 

we have gpus, you can make them go brrrr

When I have multiple Anneals and some CPT runs


Cody Blakeney 님이 재게시함

not your weights not your model not your mind

Am I wrong in sensing a paradigm shift in AI? Feels like we’re moving from a world obsessed with generalist LLM APIs to one where more and more companies are training, optimizing, and running their own models built on open source (especially smaller, specialized ones) Some…

ClementDelangue's tweet image. Am I wrong in sensing a paradigm shift in AI?

Feels like we’re moving from a world obsessed with generalist LLM APIs to one where more and more companies are training, optimizing, and running their own models built on open source (especially smaller, specialized ones)

Some…


Give it a year or two and we will have nanoagent. Then nanoagent speed run. Then nanoASI and so on.

Excited to release new repo: nanochat! (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…

karpathy's tweet image. Excited to release new repo: nanochat!
(it's among the most unhinged I've written).

Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…


Still thinking about this banger

when you're the eval guy and you check up on the experiments channel in slack

jecdohmann's tweet image. when you're the eval guy and you check up on the experiments channel in slack


Cody Blakeney 님이 재게시함

HAHAHAHAHAHA IT'S A VIDEO FROM THE PORT OF LONG BEACH CALIFORNIA 🇺🇸🇺🇸🇺🇸🇺🇸🇺🇸

🇨🇳#China’s intelligent port operates with high efficiency At China’s smart port, autonomous transport vehicles move in an orderly and tireless manner.



Its that time again

code_star's tweet image. Its that time again

Cody Blakeney 님이 재게시함

Here’s a scaling law that matters more: every two years the gap to frontier closes by a factor of 2x

Excited to release new repo: nanochat! (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…

karpathy's tweet image. Excited to release new repo: nanochat!
(it's among the most unhinged I've written).

Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single,…


I’m fairly convinced the first rabbit in a hat trick was an accident


code_star's tweet image.
code_star's tweet image.

New website looking pretty slick 👀

code_star's tweet image. New website looking pretty slick 👀


code_star's tweet image.

Talk from Wenting Zhao of Qwen on their plans during COLM. Seems like 1 word is the plan still: scaling training up! Let’s go.

natolambert's tweet image. Talk from Wenting Zhao of Qwen on their plans during COLM. Seems like 1 word is the plan still: scaling training up! Let’s go.


Doing some on the ground research on power (football) analysis

Was watching the Georgia game and noticed @dwarkesh_sp, Host of @dwarkeshpodcast was literally on TV 🤯🤯🤯🤯🤯



Cody Blakeney 님이 재게시함

One day, @code_star will swap profile pics with @willccbb and the world will tremble.

One thing I wonder is, these GPUs are going to remain incredibly powerful and the next generation or two of GPUs are going to be rolled at scales we have never really seen before. Now obviously electricity isn't free, but if we find ourselves with much cheaper electricity in…



Quit hoggin'em all

Each dot represents 5,000 hogs

uncledoomer's tweet image. Each dot represents 5,000 hogs


One thing I wonder is, these GPUs are going to remain incredibly powerful and the next generation or two of GPUs are going to be rolled at scales we have never really seen before. Now obviously electricity isn't free, but if we find ourselves with much cheaper electricity in…

So in like 5 ish years ... what is going to happen to all the H100s ... landfill?



what could this possibly mean ...

code_star's tweet image. what could this possibly mean ...

Loading...

Something went wrong.


Something went wrong.