therealokai's profile picture. Busy bee with a monkey mind.

omar khaled

@therealokai

Busy bee with a monkey mind.

置頂

After almost 3 years, I have finally learned the basics of the hottest programming language. Okay, now what's the next step to pursue a master's degree?

therealokai's tweet image. After almost 3 years, I have finally learned the basics of the hottest programming language. 
Okay, now what's the next step to pursue a master's degree?

The hottest new programming language is English



omar khaled 已轉發

رَبُّكُمْ أَعْلَمُ بِمَا فِي نُفُوسِكُمْ ۚ إِنْ تَكُونُوا صَالِحِينَ فَإِنَّهُ كَانَ لِلْأَوَّابِينَ غَفُورًا


omar khaled 已轉發

This paper asks when LLMs can be trusted to judge mental health replies. Found that LLMs systematically overrate replies, especially on empathy and helpfulness. Even when the ranking order matched human experts, the actual scores were too high, which means models look better…

rohanpaul_ai's tweet image. This paper asks when LLMs can be trusted to judge mental health replies. 

Found that LLMs systematically overrate replies, especially on empathy and helpfulness.

Even when the ranking order matched human experts, the actual scores were too high, which means models look better…

omar khaled 已轉發

مش عارفه كنت عايشة ازاي من غير ما اذاكر System Design كل ما الواحد يعرف اكتر في ال software engineering يحس ان صناعة ال software ممتعة ومعقدة مش شوية كود الواحد يكتبهم ويلزقهم مع بعض


omar khaled 已轉發

🔥 GPT-6 may not just be smarter, it might be alive (in the computational sense). A new research paper called SEAL, Self-Adapting Language Models (arXiv:2506.10943) describes how an AI can continuously learn after deployment, evolving its own internal representations without…

VraserX's tweet image. 🔥 GPT-6 may not just be smarter, it might be alive (in the computational sense).

A new research paper called SEAL, Self-Adapting Language Models (arXiv:2506.10943) describes how an AI can continuously learn after deployment, evolving its own internal representations without…

omar khaled 已轉發

Check out their blogs if you are into AI/ML. 1) Andrej Karpathy Neural networks & LLMs explained from first principles by one of the OGs of modern AI. - karpathy.ai/?utm_source=li… 2) Sebastian Raschka, PhD Deep dives into LLM training and fine-tuning with super clear code…

techwith_ram's tweet image. Check out their blogs if you are into AI/ML.

1) Andrej Karpathy 
Neural networks & LLMs explained from first principles by one of the OGs of modern AI.

- karpathy.ai/?utm_source=li…

2) Sebastian Raschka, PhD
Deep dives into LLM training and fine-tuning with super clear code…

omar khaled 已轉發

RIP fine-tuning ☠️ This new Stanford paper just killed it. It’s called 'Agentic Context Engineering (ACE)' and it proves you can make models smarter without touching a single weight. Instead of retraining, ACE evolves the context itself. The model writes, reflects, and edits…

rryssf_'s tweet image. RIP fine-tuning ☠️

This new Stanford paper just killed it.

It’s called 'Agentic Context Engineering (ACE)' and it proves you can make models smarter without touching a single weight.

Instead of retraining, ACE evolves the context itself.

The model writes, reflects, and edits…

omar khaled 已轉發

- build an autograd engine from scratch - write a mini-GPT from scratch - implement LoRA and fine-tune a model on real data - hate CUDA at least once - cry - keep going the roadmap - 5 phases - if you already know something? skip - if you're lost? rewatch - if you’re stuck? use…


omar khaled 已轉發

PPO vs GRPO vs REINFORCE – a workflow breakdown of the most talked-about reinforcement learning algorithms ➡️ Proximal Policy Optimization (PPO): The Stable Learner It’s used everywhere from dialogue agents to instruction tuning as it balances between learning fast and staying…

TheTuringPost's tweet image. PPO vs GRPO vs REINFORCE – a workflow breakdown of the most talked-about reinforcement learning algorithms

➡️ Proximal Policy Optimization (PPO): The Stable Learner

It’s used everywhere from dialogue agents to instruction tuning as it balances between learning fast and staying…

omar khaled 已轉發

Ironically, LLM researchers know nothing about language.


omar khaled 已轉發

This one paper might kill the AI scaling hype. While Big Tech burns billions on massive datasets, researchers just achieved state-of-the-art agent performance using 78 samples. And it makes a scary amount of sense. Here's the full breakdown:

connordavis_ai's tweet image. This one paper might kill the AI scaling hype.

While Big Tech burns billions on massive datasets, researchers just achieved state-of-the-art agent performance using 78 samples.

And it makes a scary amount of sense.

Here's the full breakdown:

omar khaled 已轉發

أكتوبر شهر الوقوع في الحب أنا في أكتوبر من ١٩٩٨:

2alaaaaaaaaaa2's tweet image. أكتوبر شهر الوقوع في الحب 

أنا في أكتوبر من ١٩٩٨:

omar khaled 已轉發

the closer I get to my MVP being done, the more I don’t want to ship it lol


بفتكر ايام ثانوية عامة ( و اول سنتين فالكلية) لما كان جزء من المدرسين و معيدين الكلية يعاملوني معاملة غريبة فشخ و لما اسألهم يقولولي "هو كدا شكلك مش عاجبني" لحد ما قررت استلسم و احلق

نائب رئيس الجامعه و هو بيسلمني شهاده تقدير عشان اتأهلنا لل ACPC من يومين بدل ما يقولي مبروك بيقولي ينفع شعرك دا و انا اتقتلت ضحك من بعدها قدامهم و حتي الصوره الي منزلنها باين فيها اني كنت مقتول ضحك



حاجة كدا زي "عشان تعرف حسابك قوي ولا لا اكتب الباسورد في كومنت لو ظهر **** يبقي حسابك قوي"

vibe coders, secure your systems with this tool in 1 easy step

TheAhmadOsman's tweet image. vibe coders, secure your systems with this tool in 1 easy step


omar khaled 已轉發

This Tencent paper shows a way to improve reasoning by training only on raw text using reinforcement learning. It is called Reinforcement Learning on Pre-Training data (RLPT) and it removes the need for human labels. Simple “predict the next segment” rewards are enough to…

rohanpaul_ai's tweet image. This Tencent paper shows a way to improve reasoning by training only on raw text using reinforcement learning. 

It is called Reinforcement Learning on Pre-Training data (RLPT) and it removes the need for human labels.

Simple “predict the next segment” rewards are enough to…

"Rule number one, DON'T FKN SAY 'you are absolutely right!' and write code!"

I’m tired of being “absolutely right!” when coding with an agent



omar khaled 已轉發

DSPy and ColBERT are interesting academic experiments imo. Each is a multi-paper repo that has one coherent artifact, combining our latest research together. We typically release the features as open source—hence get users/feedback—well before writing a paper on the new ideas.

Oh yes, I meant that I thought a paper came out with the thread you made about DSPy in August.



omar khaled 已轉發

Talking to grad students, too many think that long-term projects (not scattered papers), proper code releases, thoughtful benchmarks are "not incentivized". Most often they're mistaken. If we're talking incentives, *nothing* matches demonstrating impact! Will blog on this soon.


omar khaled 已轉發

Science of RL optimization is likely humanity’s last open scientific problem


omar khaled 已轉發

Prediction: In ~3 years academia will be the most desirable place to do fundamental AI research Contributing factors: - small models improve/become significantly more impactful - open weights community broadens its reach - gpus continue to get faster & cheaper - meaningful…


United States 趨勢

Loading...

Something went wrong.


Something went wrong.