
Vivek
@cuda_optimized
powerful dreamer @iiscbangalore Interests : ai, tech, f1, cricket, music
قد يعجبك
winter arc #3 (8.5hrs): -> read deepseek-r1 paper -> finished most of chapters in the rlhfbook -> watched @hkproj vid on rlhf -> contemplated some future paths -> opinionated rl blog @jsuarez5341 -> went through cublas docs -> watched @elliotarledge tensor core vids

does anyone actually use linkedIn, or do we all just log in once a month to accept random connection requests from strangers.
open source gives ideas. closed source takes them, scales them, hides them. fair game or just pure cheating?
highlight your notes in a quick and easy way. credit : @adivekar_ -> green - quick read -> yellow - read slowly and imp -> red - read, think and understand

yeah, now this makes sense.
winter arc #2 (9hrs): -> read deepseek-math paper and grpo -> finished first 8 chap of rlhfbook by @natolambert -> read @kipperrii transformer inference arithmetic -> watched @elliotarledge vid on cublas and cublasLt -> wrote sgemm & hgemm in cublas -> @karpathy nanochat

does kl penalty and grad norm ultimately have the same effect on the grpo loss. if so, then why can't we just add grad norm instead of the kl penalty term.
notes on deepseek-math paper - deepseek-math-base -> pretrained model on code and math data - deepseekmath-instruct 7B -> sft using coT, poT and tool reasoning - deepseekmath-rl -> grpo on gsm8k and math questions - rl is increasing prob of correct response

winter arc #1 (9.5hrs): -> read context & sequence parallel -> failed to impl ring attn -> binge watched @willccbb vids on yt. -> went deep into deepseek - r1 and watched some vids. -> posted tweet an gpt2 impl in triton and got a like from karpathy -> overall not a bad day!!
i love how @dwarkesh_sp is trying to convince richard sutton that next token prediction is kinda like rl
just saw the @elliotarledge yt latest vid. man that’s so deep and thoughtful on how you spoke about your highs and lows. just wanted to say your an absolute inspiration man. good things will definitely happen soon brother!! keep inspiring us with your work and time lapses!!
why chatgpt is better than google ->compression : quick answers + stores a lot information. compression ratio is very good. ->context : able to identify your problems/questions which are not there on the internet and answer specifically.
man these llm's are so good without any context i wonder what happens if we give the right context to these llm's
this is not what i expected for humans vs robots to be
United States الاتجاهات
- 1. D’Angelo 285K posts
- 2. Pentagon 107K posts
- 3. Brown Sugar 20.6K posts
- 4. #PortfolioDay 16.2K posts
- 5. Young Republicans 13.8K posts
- 6. Politico 164K posts
- 7. Drew Struzan 28.3K posts
- 8. Big 12 13.2K posts
- 9. Scream 5 N/A
- 10. Black Messiah 10.8K posts
- 11. David Bell N/A
- 12. Milei 262K posts
- 13. Presidential Medal of Freedom 59.3K posts
- 14. Soybeans 5,228 posts
- 15. Merino 14.7K posts
- 16. Venables 3,641 posts
- 17. Nick Mangold N/A
- 18. World Cup 334K posts
- 19. Voodoo 21.5K posts
- 20. Baldwin 20.9K posts
قد يعجبك
-
dill
@dill_sunnyb11 -
Bruno Henrique
@Brunot3ch -
Salman Ibne Eunus
@ibne_eunus -
Martin Andrews
@mdda123 -
T Tian
@gdsttian -
Bertrand Couture
@bertrandcouture -
Mauricio Alzate
@Xarahenergy -
AMVITABLE®
@amvitable -
Elias
@notes_own -
Milind Hanchinmani
@mhanchinmani -
GabrielOrtega
@Tsunami70510954 -
Sophy Sem
@semsphy -
Anuj Dutt
@anujdutt92 -
DOKON🚃
@dokondokon -
Aditya Morolia
@AdityaMorolia
Something went wrong.
Something went wrong.