You might like
Stanford just dropped a 1:47 hour masterclass about LLM training on YouTube. this is the 4th lecture in their new Fall 2025 series of transformers and LLM. this is the new go-to course on LLMs imo, and you need to take it seriously. subjects in this video: → Pretraining →…
Dear Telecom co, I would like to share two humble suggestions: Many Indians today already have Wi-Fi connections at home, especially senior citizens who primarily rely on it. It would be great if COs could introduce a low-cost plan offering around 1 GB per day, with flexible…
This video explains how diffusion models are overtaking Large Language Models for generation tasks like: 1. Code Generation 2. Image Generation 3. Video Generation 00:00 Agenda 00:20 How are they different from LLMs? 05:09 Internal Mechanism 10:09 How are vectors generated?…
Amazon just laid off 14000 people. Not a single minister has offered any support. > Govt collects taxes from IT professionals > Use same taxes to fund freebies > Use those freebies to win elections But when these taxpayers are in trouble, government is nowhere to be found.
An exciting new professional certificate: PyTorch for Deep Learning taught by @lmoroney is now available at DeepLearning.AI. This is the definitive program for learning PyTorch, which is one of the main frameworks researchers use to build breakthrough AI systems. If you…
krupadave.com
Everything About Transformers
Illustrated guide to how transformers actually work — and why they’re built this way.
🔗 Mixture of Experts (MoE): github.com/rasbt/LLMs-fro…
Announcing my new course: Agentic AI! Building AI agents is one of the most in-demand skills in the job market. This course, available now at deeplearning.ai, teaches you how. You'll learn to implement four key agentic design patterns: - Reflection, in which an agent…
Our course recommendation of the day is “Post-training of LLMs, ” where you’ll learn how to customize pre-trained language models using Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO), and Online Reinforcement Learning (RL). You'll learn when to use each…
Unsloth now has a Docker image! 🐳 Train LLMs locally with no setup: just run the image and go. Includes every pre-made Unsloth notebook. Solves dependency or environment issues. Guide: docs.unsloth.ai/new/how-to-tra…
still experimenting with LoRA based on the @thinkymachines configuration and just implemented it in colab. In this notebook I set up a fine tune of Qwen/Qwen3-0.6B on the OpenR1-Math dataset with lora rank of 1. with this setup you can get the same reward accuracy as full…
Introducing Tinker: a flexible API for fine-tuning language models. Write training loops in Python on your laptop; we'll run them on distributed GPUs. Private beta starts today. We can't wait to see what researchers and developers build with cutting-edge open models!…
Announcing our first product: Tinker! Tinker is a training API for everyone! It lets you focus on what matters in LLM training - your data and algorithms - while we handle the heavy lifting of distributed training. You can train your own models using Tinker even if you have no…
Introducing Tinker: a flexible API for fine-tuning language models. Write training loops in Python on your laptop; we'll run them on distributed GPUs. Private beta starts today. We can't wait to see what researchers and developers build with cutting-edge open models!…
Red carpet welcome at @Rakuten India office in Bengaluru! Kaadubesanahalli, Prestige Tech Park! Global companies trust you and set up shop here, and this is the kind of civic amenities you offer! @GBAChiefComm and then our DCM @DKShivakumar say companies can’t BLACKMAIL… does…
Some resources about GPUs I found good as a noob in GPU programming, 1. jax-ml.github.io/scaling-book/g… 2. modal.com/gpu-glossary/p… 3. multimodalai.substack.com/p/the-mlai-eng… 4. bytesofintelligence.substack.com/p/maximizing-g… 5. youtube.com/playlist?list=…
The more charts I see the more I get convinced: A big big correction is due next year #2026. Till then enjoy the rally that's right in front of us for the next few months!! #Nifty #Nifty500 #NiftyMidcap100 #NiftySmallcap100
solid in-depth explanation of paged attention in this blog.
#Nifty 24741 7th Sep 2024 And today 7th Sep 2025 A year has passed but the story continues!! That time we were about to end the 3rd and rest is history. The 4th took over! Now, we are running that 5th and that may complete in the next 3-4 months or by very early next year.…
You can now run gpt-oss-120b & 20b locally with our GGUFs! 🦥 Run OpenAI's 120b model on 66GB RAM & 20b model on 14GB RAM. Both in original precision. Uploads includes our chat template fixes. Guide: docs.unsloth.ai/basics/gpt-oss GGUF: huggingface.co/unsloth/gpt-os…
United States Trends
- 1. Lakers 66.1K posts
- 2. Luka 63.6K posts
- 3. Wemby 24.7K posts
- 4. Marcus Smart 5,328 posts
- 5. #LakeShow 5,262 posts
- 6. Blazers 7,695 posts
- 7. Russ 9,565 posts
- 8. Ayton 14.2K posts
- 9. Richard 44.5K posts
- 10. Horford 1,798 posts
- 11. #AmphoreusStamp 5,444 posts
- 12. #RipCity N/A
- 13. Podz 2,325 posts
- 14. Champagnie 1,195 posts
- 15. Spencer Knight N/A
- 16. #AEWDynamite 20K posts
- 17. Thunder 31K posts
- 18. Kuminga 3,249 posts
- 19. Nico Harrison 1,557 posts
- 20. #dispatch 59.9K posts
Something went wrong.
Something went wrong.