Ethan
@torchcompiled
trying to feel the magic. cofounder at @leonardoai_ research at @canva
Vous pourriez aimer
personally I feel like the inflection point was early 2022. The sweet spot where clip-guided diffusion was just taking off, forcing unconditional models to be conditional through strange patchwork of CLIP evaluating slices of the canvas at a time. It was like improv, always…
Image synthesis used to look so good. These are from 2021. I feel like this was an inflection point, and the space has metastasized into something abhorrent today (Grok, etc). Even with no legible representational forms, there was so much possibility in these images.
Trying to get an LLM to solve a math problem, running through tons of matmuls and writing every token, feels like trying to bowl a strike all the way from the moon while drunk. But training kinda places the guardrails so that randomly walking bowling ball stays on track.
I like JEPA but this image is such a meme at this point
I want a good text VAE that works over long pieces of text. I would love to see a poem or something I wrote in an alternate portrayal. Apparently, this is quite difficult. Autoregressive decoding of latents quickly shifts from depending on latent variable to depending on past…
claim: the natural state of neural networks are actually incredibly unbiased, following from loss functions that are unbiased estimators of distributions. they may cling onto spurious correlations but arguably this is a different problem of generalization.
one of my favorite outputs from the guided diffusion days, a conductor and his orchestra in the middle of a storm
Little stake here as saying humans are thinking in an autoregressive or diffusion manner is overly reductive. But a fun fact is that human associative memory does share some similarities with Hopfield networks which somewhat resembles a denoising process
> The way humans think look a lot more like diffusion than autoregressive. i will never, ever understand this claim or the intuitions behind it. ah yes. the human mind is... learning a scoring function to... reverse gaussian noise... (?) ... spatially (???)
> search megatron hoping to find the training library > get the character > "megatron transformers" > still getting the character > I give up
What happens if we make the tiny reasoning model like 2B instead of 70m
United States Tendances
- 1. Good Thursday 19.7K posts
- 2. Knicks 13.7K posts
- 3. Shamet 2,985 posts
- 4. #AEWDynamite 22K posts
- 5. FEMA 70.7K posts
- 6. Sam Harris 3,408 posts
- 7. Brandon Williams 1,064 posts
- 8. #Survivor49 3,973 posts
- 9. NO CAP 15.2K posts
- 10. Derik Queen 4,632 posts
- 11. #AEWCollision 8,873 posts
- 12. #SeeRed N/A
- 13. #TheChallenge41 1,687 posts
- 14. Lute 89.1K posts
- 15. Nany 2,054 posts
- 16. Chisa 35.1K posts
- 17. Sheila Cherfilus-McCormick 53.2K posts
- 18. Blazers 4,253 posts
- 19. Josh Hart 2,889 posts
- 20. Crockett 54.3K posts
Vous pourriez aimer
-
Alexander S
@devdef -
KaliYuga
@KaliYuga_ai -
pharmapsychotic
@pharmapsychotic -
Shade.
@Shade_9S0 -
apolinario 🌐
@multimodalart -
huemin
@huemin_art -
Takyon∞
@takyon236 -
Thibaud Zamora
@thibaudz -
Nerdy Rodent 🐀🤓💻🪐🚴
@NerdyRodent -
proxima centauri b
@proximasan -
sure, ai
@sureailabs -
nin
@nin_artificial -
Chris Allen
@zippy731 -
KaptN
@Legendsartcoll1 -
|ᴬ𝖎| Joe
@AiJoe_eth
Something went wrong.
Something went wrong.