omar khaled
@therealokai
Busy bee with a monkey mind.
After almost 3 years, I have finally learned the basics of the hottest programming language. Okay, now what's the next step to pursue a master's degree?
The hottest new programming language is English
رَبُّكُمْ أَعْلَمُ بِمَا فِي نُفُوسِكُمْ ۚ إِنْ تَكُونُوا صَالِحِينَ فَإِنَّهُ كَانَ لِلْأَوَّابِينَ غَفُورًا
This paper asks when LLMs can be trusted to judge mental health replies. Found that LLMs systematically overrate replies, especially on empathy and helpfulness. Even when the ranking order matched human experts, the actual scores were too high, which means models look better…
مش عارفه كنت عايشة ازاي من غير ما اذاكر System Design كل ما الواحد يعرف اكتر في ال software engineering يحس ان صناعة ال software ممتعة ومعقدة مش شوية كود الواحد يكتبهم ويلزقهم مع بعض
🔥 GPT-6 may not just be smarter, it might be alive (in the computational sense). A new research paper called SEAL, Self-Adapting Language Models (arXiv:2506.10943) describes how an AI can continuously learn after deployment, evolving its own internal representations without…
Check out their blogs if you are into AI/ML. 1) Andrej Karpathy Neural networks & LLMs explained from first principles by one of the OGs of modern AI. - karpathy.ai/?utm_source=li… 2) Sebastian Raschka, PhD Deep dives into LLM training and fine-tuning with super clear code…
RIP fine-tuning ☠️ This new Stanford paper just killed it. It’s called 'Agentic Context Engineering (ACE)' and it proves you can make models smarter without touching a single weight. Instead of retraining, ACE evolves the context itself. The model writes, reflects, and edits…
- build an autograd engine from scratch - write a mini-GPT from scratch - implement LoRA and fine-tune a model on real data - hate CUDA at least once - cry - keep going the roadmap - 5 phases - if you already know something? skip - if you're lost? rewatch - if you’re stuck? use…
PPO vs GRPO vs REINFORCE – a workflow breakdown of the most talked-about reinforcement learning algorithms ➡️ Proximal Policy Optimization (PPO): The Stable Learner It’s used everywhere from dialogue agents to instruction tuning as it balances between learning fast and staying…
Ironically, LLM researchers know nothing about language.
This one paper might kill the AI scaling hype. While Big Tech burns billions on massive datasets, researchers just achieved state-of-the-art agent performance using 78 samples. And it makes a scary amount of sense. Here's the full breakdown:
أكتوبر شهر الوقوع في الحب أنا في أكتوبر من ١٩٩٨:
the closer I get to my MVP being done, the more I don’t want to ship it lol
بفتكر ايام ثانوية عامة ( و اول سنتين فالكلية) لما كان جزء من المدرسين و معيدين الكلية يعاملوني معاملة غريبة فشخ و لما اسألهم يقولولي "هو كدا شكلك مش عاجبني" لحد ما قررت استلسم و احلق
نائب رئيس الجامعه و هو بيسلمني شهاده تقدير عشان اتأهلنا لل ACPC من يومين بدل ما يقولي مبروك بيقولي ينفع شعرك دا و انا اتقتلت ضحك من بعدها قدامهم و حتي الصوره الي منزلنها باين فيها اني كنت مقتول ضحك
حاجة كدا زي "عشان تعرف حسابك قوي ولا لا اكتب الباسورد في كومنت لو ظهر **** يبقي حسابك قوي"
vibe coders, secure your systems with this tool in 1 easy step
This Tencent paper shows a way to improve reasoning by training only on raw text using reinforcement learning. It is called Reinforcement Learning on Pre-Training data (RLPT) and it removes the need for human labels. Simple “predict the next segment” rewards are enough to…
"Rule number one, DON'T FKN SAY 'you are absolutely right!' and write code!"
I’m tired of being “absolutely right!” when coding with an agent
DSPy and ColBERT are interesting academic experiments imo. Each is a multi-paper repo that has one coherent artifact, combining our latest research together. We typically release the features as open source—hence get users/feedback—well before writing a paper on the new ideas.
Oh yes, I meant that I thought a paper came out with the thread you made about DSPy in August.
Talking to grad students, too many think that long-term projects (not scattered papers), proper code releases, thoughtful benchmarks are "not incentivized". Most often they're mistaken. If we're talking incentives, *nothing* matches demonstrating impact! Will blog on this soon.
Science of RL optimization is likely humanity’s last open scientific problem
Prediction: In ~3 years academia will be the most desirable place to do fundamental AI research Contributing factors: - small models improve/become significantly more impactful - open weights community broadens its reach - gpus continue to get faster & cheaper - meaningful…
United States Trends
- 1. #SmackDown 41.1K posts
- 2. Norvell 2,954 posts
- 3. Mamdani 412K posts
- 4. Reed Sheppard 2,763 posts
- 5. Florida State 10.7K posts
- 6. Marjorie Taylor Greene 55.2K posts
- 7. NC State 5,326 posts
- 8. #OPLive 2,387 posts
- 9. Collin Gillespie 1,716 posts
- 10. Sengun 6,713 posts
- 11. Suns 12.1K posts
- 12. #BostonBlue 3,742 posts
- 13. Wolves 16.2K posts
- 14. Syla Swords 2,959 posts
- 15. Booker 7,029 posts
- 16. Dillon Brooks 2,515 posts
- 17. Anthony Edwards 2,448 posts
- 18. Azzi 15.8K posts
- 19. Derik Queen 4,772 posts
- 20. Rockets 15.2K posts
Something went wrong.
Something went wrong.