cuda_optimized's profile picture. powerful dreamer @iiscbangalore Interests : ai, tech, f1, cricket, music

Vivek

@cuda_optimized

powerful dreamer @iiscbangalore Interests : ai, tech, f1, cricket, music

고정된 트윗

woke up to this!! 😱😱

cuda_optimized's tweet image. woke up to this!! 😱😱

does anyone actually use linkedIn, or do we all just log in once a month to accept random connection requests from strangers.


open source gives ideas. closed source takes them, scales them, hides them. fair game or just pure cheating?


highlight your notes in a quick and easy way. credit : @adivekar_ -> green - quick read -> yellow - read slowly and imp -> red - read, think and understand

cuda_optimized's tweet image. highlight your notes in a quick and easy way. credit : @adivekar_ 
-> green - quick read
-> yellow - read slowly and imp
-> red - read, think and understand

yeah, now this makes sense.


btw, @waitin4agi_ cooked this one!!


winter arc #2 (9hrs): -> read deepseek-math paper and grpo -> finished first 8 chap of rlhfbook by @natolambert -> read @kipperrii transformer inference arithmetic -> watched @elliotarledge vid on cublas and cublasLt -> wrote sgemm & hgemm in cublas -> @karpathy nanochat

cuda_optimized's tweet image. winter arc #2 (9hrs):
-> read deepseek-math paper and grpo
-> finished first 8 chap of rlhfbook by @natolambert 
-> read @kipperrii  transformer inference arithmetic
-> watched @elliotarledge  vid on cublas and cublasLt
-> wrote sgemm & hgemm in cublas
-> @karpathy nanochat

does kl penalty and grad norm ultimately have the same effect on the grpo loss. if so, then why can't we just add grad norm instead of the kl penalty term.


notes on deepseek-math paper - deepseek-math-base -> pretrained model on code and math data - deepseekmath-instruct 7B -> sft using coT, poT and tool reasoning - deepseekmath-rl -> grpo on gsm8k and math questions - rl is increasing prob of correct response

cuda_optimized's tweet image. notes on deepseek-math paper
- deepseek-math-base -> pretrained model on code and math data
- deepseekmath-instruct 7B -> sft using coT, poT and tool reasoning
- deepseekmath-rl -> grpo on gsm8k and math questions
- rl is increasing prob of correct response

winter arc #1 (9.5hrs): -> read context & sequence parallel -> failed to impl ring attn -> binge watched @willccbb vids on yt. -> went deep into deepseek - r1 and watched some vids. -> posted tweet an gpt2 impl in triton and got a like from karpathy -> overall not a bad day!!


most of the llm's today are a simp.


i love how @dwarkesh_sp is trying to convince richard sutton that next token prediction is kinda like rl


just saw the @elliotarledge yt latest vid. man that’s so deep and thoughtful on how you spoke about your highs and lows. just wanted to say your an absolute inspiration man. good things will definitely happen soon brother!! keep inspiring us with your work and time lapses!!


why chatgpt is better than google ->compression : quick answers + stores a lot information. compression ratio is very good. ->context : able to identify your problems/questions which are not there on the internet and answer specifically.


man these llm's are so good without any context i wonder what happens if we give the right context to these llm's


this is not what i expected for humans vs robots to be

I’m sorry but WHAT THE FUCK?!



sam altman has a way of answering the question without actually answering the question while doing a podcast


not sure why but after a number of responses grok tends to repeats previous answer. @xai


Loading...

Something went wrong.


Something went wrong.