Mr. L
@JrtheProgrammer
IT #nextworld / Researcher Enthusiast #TheBoringFundamentals
Agentic AI Overview This report provides a comprehensive overview of architectures, applications, and future directions. Great read for AI devs and enthusiasts. It introduces a new dual-paradigm framework that categorizes agentic systems into two distinct lineages: the…
ChatGPT for Research and Publication: A Step-by-Step Guide If you need the PDF version, please comment below
My PhD thesis--On Zero-Shot Reinforcement Learning--is now on arXiv.
The most comprehensive, LLM architecture analysis I've read. Covers every flagship model: 1. DeepSeek V3/R1 2. OLMo 2 3. Gemma 3 4. Mistral Small 3.1 5. Llama 4 6. Qwen3 7. SmolLM3 8. Kimi 2 9. GPT-OSS Great article by @rasbt🙌 Link in the comments 👇 ♻️ Repost if you…
A free book 👇 "Foundations of Lange Language Models" by Tong Xiao and Jingbo Zhu It's good to refresh the core concepts and techniques behind LLMs. This 230-page book covers topics, such as: - Pre-training - Generative models (training, fine-tuning, memory, scaling) -…
Financial Statement Analysis with Large Language Models (LLMs) A 54-page PDF:
1/ With @BenDLaufer and Jon Kleinberg, we constructed the largest dataset of its kind to date: 1.86M Hugging Face models. In a new paper, we mapped how the open-source AI ecosystem evolves by tracing fine-tunes, merges, and more. Here's what we found 🧵
The freshest AI/ML research of the week Our top 9 ▪️ Sotopia-RL: Reward Design for Social Intelligence ▪️ Agent Lightning: Train ANY AI Agents with RL ▪️ Exploitation Is All You Need... for Exploration ▪️ Learning to Reason for Factuality ▪️ VeOmni ▪️ Is Chain-of-Thought…
Been working HRM, had been getting mixed results. AdamAtan2 usage is interesting. Paper covers Sudoku and ARC AGI 1/2. These are essentially step-based grid struct prob. Anyone working w/ HRM & finding other interesting examples? Seen tons of hype, but v few people implementing.
The importance of stupidity in scientific research
Training an LLM on 8 M4 Mac Minis Ethernet interconnect between Macs is 100x slower than NVLink so Macs can’t synchronise model gradients every training step. I got DiLoCo running so Macs synchronise once every 1000 training steps using 1000x less communication than DDP
Most people are still prompting wrong. I've found this framework, which was even shared by OpenAI President Greg Brockman. Here’s how it works:
Local Deep Research - A local LLM research assistant that generates follow-up questions and uses DuckDuckGo for web searches - Runs 100% locally with Ollama - Works with Mistral 7B or DeepSeek 14B - Generates structured research reports with sources
my roadmap to learning LLMs - electrons - circuits - logic - transistors - comp arch - CPUs - GPUs - linear algebra - probability - machine learning - optimization - optimizers - tokenization - transformers - pretraining - distributed training - RL - post training - distillation…
You can solve 80% of interview problems about strings with a basic approach. But if the question is tricky, you probably have to think about tries. Tries are unique data structures you can use to represent strings efficiently. This is how to use them: ↓
OpenAI has released a new prompting guide for their reasoning models. It emphasizes simplicity, avoiding chain-of-thought prompts, the use of delimiters, and when to use them. Here’s a breakdown and an optimized prompt to have it write like you:
Financial Statement Analysis with Large Language Models (LLMs) A 54-page PDF:
United States Trends
- 1. FIFA 128K posts
- 2. FINALLY DID IT 384K posts
- 3. The WET 107K posts
- 4. $MAYHEM 2,689 posts
- 5. Infantino 20.7K posts
- 6. The BONK 241K posts
- 7. Warner Bros 178K posts
- 8. The Jupiter 52.8K posts
- 9. #NXXT_AI_Energy N/A
- 10. #FanCashDropPromotion 2,521 posts
- 11. HBO Max 73K posts
- 12. Kevin Hart 3,348 posts
- 13. #NXXT_CleanEra N/A
- 14. NextNRG Inc 2,861 posts
- 15. Chris Henry Jr 9,162 posts
- 16. Hep B 1,361 posts
- 17. #GenerationsShift_NXXT N/A
- 18. Andrea Bocelli 4,252 posts
- 19. Robbie Williams 2,528 posts
- 20. Paramount 31.4K posts
Something went wrong.
Something went wrong.