내가 좋아할 만한 콘텐츠
I implemented gpt-2 using graphics shaders
vibecoded a visualizer for latent representations of tokens as they flow through gpt-2 (runs entirely in browser on webgl) visualizing high dim space is hard so rather than just PCAing 768 -> 2, it renders similarity as graph connectedness at an adjustable cosine threshold
xAI should fork VS Code and drop a cursor competitor. They could name it Xcode
I’m surprised that no app uses face verification (like what dating apps have) to combat bots. Deleting an account is terrible for false positives, but having to do face verification or be locked out of your account is just an inconvenience. Would solve most of the issue imo
My experiments with fine-tuning RoBERTa to do language diffusion are at an end. Surprisingly cohesive with such a minimum implementation but not as good as gpt-2. A more thorough implementation (and better training) should be able to reach parity on quality and speed though.
Just read how Fourier transforms “work” because sines and cosines form an orthogonal basis for a specific Hilbert space of functions. Math is beautiful but it always feels like a bottomless pit of knowledge where there’s always an infinite amount of things you don’t know
I don’t know why LLM companies don’t watermark their output by using rare UTF-8 code points for similar looking characters. If they just replaced all U+002D: HYPHEN-MINUS with U+2010: HYPHEN, basically no one would notice but it’d be obvious to software that it’s generated output

I was not put on this earth to use python libraries all day
In the early days of X/PayPal, fraud was their biggest problem and they built out tools for anomaly detection which eventually led to Palantir. It’s ironic that 20 years later, the reincarnation of X (via twitter) does such a bad job of anomaly (bot account) detection
Increased the number of diffusion steps for my RoBERTa Diffusion model and it’s wild how surprisingly good this is. Will fine-tune it on OpenWebText and compare it to GPT-2 later
I’m am surprised at the amount of coherency I’ve gotten by trying to fine-tune RoBERTa into a language diffusion model. Pretty decent for a 6 year-old model with only 125 million parameters
New Paper! Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents A longstanding goal of AI research has been the creation of AI that can learn indefinitely. One path toward that goal is an AI that improves itself by rewriting its own code, including any code…
Day 1 of messing with masked language diffusion models 🫠🫠
Don't use structured output mode for reasoning tasks. We’re open sourcing Osmosis-Structure-0.6B: an extremely small model that can turn any unstructured data into any format (e.g. JSON schema). Use it with any model - download and blog below!
When I started using Arch Linux years ago, any time something would randomly break I’d have to spend at least an hour sifting through forums to find a solution. Now ChatGPT can diagnose and fix it in a few seconds and Pewdiepie uses Arch
United States 트렌드
- 1. Yamamoto 48.2K posts
- 2. #DWTS 44.3K posts
- 3. halsey 9,084 posts
- 4. Growth Path 1,638 posts
- 5. #FlyTogether 2,895 posts
- 6. Young Republicans 76.4K posts
- 7. Kreider 1,307 posts
- 8. Ohtani 14.2K posts
- 9. Jared Butler N/A
- 10. #TexasHockey 3,421 posts
- 11. #WWENXT 20.3K posts
- 12. #MOST_WANTED_IN_CHICAGO 1,201 posts
- 13. TOKYO NARITA N/A
- 14. Jarry N/A
- 15. Will Richard 2,660 posts
- 16. Cuffem 2,578 posts
- 17. Tami 4,562 posts
- 18. Lucia 63.2K posts
- 19. bush 40.4K posts
- 20. Ayton 2,505 posts
내가 좋아할 만한 콘텐츠
-
ishan
@0xishand -
Brian
@TheAustinIPGuy -
Hongpeng Jin
@HongpengJin -
Bhasker Sri Harsha
@BhaskerSriHarsh -
Andy M.
@omgwtfitsp -
Edgars Liepa
@liepa_edgars -
Rich Wyatt - Author
@ManTheForce -
Pranav Mishra
@navamishra -
V. Pan
@AI_ML_iQ -
Alessandro Scarcella
@__alesca__ -
peter chung
@peterkchung -
Madysn Watkins
@MadysnWatkins
Something went wrong.
Something went wrong.