내가 좋아할 만한 콘텐츠
does anyone actually use linkedIn, or do we all just log in once a month to accept random connection requests from strangers.
open source gives ideas. closed source takes them, scales them, hides them. fair game or just pure cheating?
highlight your notes in a quick and easy way. credit : @adivekar_ -> green - quick read -> yellow - read slowly and imp -> red - read, think and understand

yeah, now this makes sense.
winter arc #2 (9hrs): -> read deepseek-math paper and grpo -> finished first 8 chap of rlhfbook by @natolambert -> read @kipperrii transformer inference arithmetic -> watched @elliotarledge vid on cublas and cublasLt -> wrote sgemm & hgemm in cublas -> @karpathy nanochat

does kl penalty and grad norm ultimately have the same effect on the grpo loss. if so, then why can't we just add grad norm instead of the kl penalty term.
notes on deepseek-math paper - deepseek-math-base -> pretrained model on code and math data - deepseekmath-instruct 7B -> sft using coT, poT and tool reasoning - deepseekmath-rl -> grpo on gsm8k and math questions - rl is increasing prob of correct response

winter arc #1 (9.5hrs): -> read context & sequence parallel -> failed to impl ring attn -> binge watched @willccbb vids on yt. -> went deep into deepseek - r1 and watched some vids. -> posted tweet an gpt2 impl in triton and got a like from karpathy -> overall not a bad day!!
i love how @dwarkesh_sp is trying to convince richard sutton that next token prediction is kinda like rl
just saw the @elliotarledge yt latest vid. man that’s so deep and thoughtful on how you spoke about your highs and lows. just wanted to say your an absolute inspiration man. good things will definitely happen soon brother!! keep inspiring us with your work and time lapses!!
why chatgpt is better than google ->compression : quick answers + stores a lot information. compression ratio is very good. ->context : able to identify your problems/questions which are not there on the internet and answer specifically.
man these llm's are so good without any context i wonder what happens if we give the right context to these llm's
this is not what i expected for humans vs robots to be
sam altman has a way of answering the question without actually answering the question while doing a podcast
United States 트렌드
- 1. Happy Birthday Charlie 50.6K posts
- 2. Good Tuesday 31.4K posts
- 3. #tuesdayvibe 3,390 posts
- 4. Shilo 1,457 posts
- 5. #NationalDessertDay N/A
- 6. #PutThatInYourPipe N/A
- 7. Pentagon 74.8K posts
- 8. Standard Time 2,995 posts
- 9. #TacoTuesday N/A
- 10. #Worlds2025 43.2K posts
- 11. Happy 32nd 8,932 posts
- 12. Victory Tuesday N/A
- 13. Dissidia 6,543 posts
- 14. Happy Birthday in Heaven 1,489 posts
- 15. Martin Sheen 6,473 posts
- 16. Janet Mills 1,561 posts
- 17. No American 64.3K posts
- 18. JPMorgan 12.3K posts
- 19. Monad 190K posts
- 20. Presidential Medal of Freedom 29.7K posts
내가 좋아할 만한 콘텐츠
-
dill
@dill_sunnyb11 -
Bruno Henrique
@Brunot3ch -
Salman Ibne Eunus
@ibne_eunus -
Martin Andrews
@mdda123 -
T Tian
@gdsttian -
Bertrand Couture
@bertrandcouture -
Mauricio Alzate
@Xarahenergy -
AMVITABLE®
@amvitable -
Elias
@notes_own -
Milind Hanchinmani
@mhanchinmani -
GabrielOrtega
@Tsunami70510954 -
Sophy Sem
@semsphy -
Anuj Dutt
@anujdutt92 -
DOKON🚃
@dokondokon -
Aditya Morolia
@AdityaMorolia
Something went wrong.
Something went wrong.