JrtheProgrammer's profile picture. IT #nextworld / Researcher Enthusiast
#TheBoringFundamentals

Mr. L

@JrtheProgrammer

IT #nextworld / Researcher Enthusiast #TheBoringFundamentals

Mr. L 님이 재게시함

My PhD thesis--On Zero-Shot Reinforcement Learning--is now on arXiv.

enjeeneer's tweet image. My PhD thesis--On Zero-Shot Reinforcement Learning--is now on arXiv.

Mr. L 님이 재게시함

The most comprehensive, LLM architecture analysis I've read. Covers every flagship model: 1. DeepSeek V3/R1 2. OLMo 2 3. Gemma 3 4. Mistral Small 3.1 5. Llama 4 6. Qwen3 7. SmolLM3 8. Kimi 2 9. GPT-OSS Great article by @rasbt🙌 Link in the comments 👇 ♻️ Repost if you…

_rohit_tiwari_'s tweet image. The most comprehensive,

LLM architecture analysis I've read.

Covers every flagship model:

1. DeepSeek V3/R1
2. OLMo 2
3. Gemma 3
4. Mistral Small 3.1
5. Llama 4
6. Qwen3
7. SmolLM3
8. Kimi 2 
9. GPT-OSS

Great article by @rasbt🙌

Link in the comments 👇

♻️ Repost if you…

Mr. L 님이 재게시함

A free book 👇 "Foundations of Lange Language Models" by Tong Xiao and Jingbo Zhu It's good to refresh the core concepts and techniques behind LLMs. This 230-page book covers topics, such as: - Pre-training - Generative models (training, fine-tuning, memory, scaling) -…

TheTuringPost's tweet image. A free book 👇

"Foundations of Lange Language Models" by Tong Xiao and Jingbo Zhu

It's good to refresh the core concepts and techniques behind LLMs.

This 230-page book covers topics, such as:

- Pre-training
- Generative models (training, fine-tuning, memory, scaling)
-…
TheTuringPost's tweet image. A free book 👇

"Foundations of Lange Language Models" by Tong Xiao and Jingbo Zhu

It's good to refresh the core concepts and techniques behind LLMs.

This 230-page book covers topics, such as:

- Pre-training
- Generative models (training, fine-tuning, memory, scaling)
-…

Mr. L 님이 재게시함

Financial Statement Analysis with Large Language Models (LLMs) A 54-page PDF:

quantscience_'s tweet image. Financial Statement Analysis with Large Language Models (LLMs)

A 54-page PDF:

Mr. L 님이 재게시함

1/ With @BenDLaufer and Jon Kleinberg, we constructed the largest dataset of its kind to date: 1.86M Hugging Face models. In a new paper, we mapped how the open-source AI ecosystem evolves by tracing fine-tunes, merges, and more. Here's what we found 🧵

didaoh's tweet image. 1/ With @BenDLaufer and Jon Kleinberg, we constructed the largest dataset of its kind to date: 1.86M Hugging Face models. In a new paper, we mapped how the open-source AI ecosystem evolves by tracing fine-tunes, merges, and more. Here's what we found 🧵

Mr. L 님이 재게시함

The freshest AI/ML research of the week Our top 9 ▪️ Sotopia-RL: Reward Design for Social Intelligence ▪️ Agent Lightning: Train ANY AI Agents with RL ▪️ Exploitation Is All You Need... for Exploration ▪️ Learning to Reason for Factuality ▪️ VeOmni ▪️ Is Chain-of-Thought…

TheTuringPost's tweet image. The freshest AI/ML research of the week

Our top 9
▪️ Sotopia-RL: Reward Design for Social Intelligence
▪️ Agent Lightning: Train ANY AI Agents with RL
▪️ Exploitation Is All You Need... for Exploration
▪️ Learning to Reason for Factuality
▪️ VeOmni
▪️ Is Chain-of-Thought…

Mr. L 님이 재게시함

Been working HRM, had been getting mixed results. AdamAtan2 usage is interesting. Paper covers Sudoku and ARC AGI 1/2. These are essentially step-based grid struct prob. Anyone working w/ HRM & finding other interesting examples? Seen tons of hype, but v few people implementing.

Sentdex's tweet image. Been working HRM, had been getting mixed results. AdamAtan2 usage is interesting. Paper covers Sudoku and ARC AGI 1/2. These are essentially step-based grid struct prob. Anyone working w/ HRM & finding other interesting examples? Seen tons of hype, but v few people implementing.

Mr. L 님이 재게시함

The importance of stupidity in scientific research

ScholarshipfPhd's tweet image. The importance of stupidity in scientific research

Mr. L 님이 재게시함

Training an LLM on 8 M4 Mac Minis Ethernet interconnect between Macs is 100x slower than NVLink so Macs can’t synchronise model gradients every training step. I got DiLoCo running so Macs synchronise once every 1000 training steps using 1000x less communication than DDP

MattBeton's tweet image. Training an LLM on 8 M4 Mac Minis

Ethernet interconnect between Macs is 100x slower than NVLink so Macs can’t synchronise model gradients every training step.

I got DiLoCo running so Macs synchronise once every 1000 training steps using 1000x less communication than DDP

Mr. L 님이 재게시함

Most people are still prompting wrong. I've found this framework, which was even shared by OpenAI President Greg Brockman. Here’s how it works:

aakashg0's tweet image. Most people are still prompting wrong.

I've found this framework, which was even shared by OpenAI President Greg Brockman.

Here’s how it works:

Mr. L 님이 재게시함

Local Deep Research - A local LLM research assistant that generates follow-up questions and uses DuckDuckGo for web searches - Runs 100% locally with Ollama - Works with Mistral 7B or DeepSeek 14B - Generates structured research reports with sources


Mr. L 님이 재게시함

my roadmap to learning LLMs - electrons - circuits - logic - transistors - comp arch - CPUs - GPUs - linear algebra - probability - machine learning - optimization - optimizers - tokenization - transformers - pretraining - distributed training - RL - post training - distillation…


Mr. L 님이 재게시함

You can solve 80% of interview problems about strings with a basic approach. But if the question is tricky, you probably have to think about tries. Tries are unique data structures you can use to represent strings efficiently. This is how to use them: ↓

Franc0Fernand0's tweet image. You can solve 80% of interview problems about strings with a basic approach.

But if the question is tricky, you probably have to think about tries.

Tries are unique data structures you can use to represent strings efficiently.

This is how to use them: ↓

Mr. L 님이 재게시함

Small teams are the future:

benln's tweet image. Small teams are the future:

Mr. L 님이 재게시함

OpenAI has released a new prompting guide for their reasoning models. It emphasizes simplicity, avoiding chain-of-thought prompts, the use of delimiters, and when to use them. Here’s a breakdown and an optimized prompt to have it write like you:

dr_cintas's tweet image. OpenAI has released a new prompting guide for their reasoning models.

It emphasizes simplicity, avoiding chain-of-thought prompts, the use of delimiters, and when to use them.

Here’s a breakdown and an optimized prompt to have it write like you:
dr_cintas's tweet image. OpenAI has released a new prompting guide for their reasoning models.

It emphasizes simplicity, avoiding chain-of-thought prompts, the use of delimiters, and when to use them.

Here’s a breakdown and an optimized prompt to have it write like you:

Mr. L 님이 재게시함

GPU-accelerated NumPy/SciPy library

tom_doerr's tweet image. GPU-accelerated NumPy/SciPy library

Mr. L 님이 재게시함

Machine Learning in C

chessMan786's tweet image. Machine Learning in C

Mr. L 님이 재게시함

Financial Statement Analysis with Large Language Models (LLMs) A 54-page PDF:

quantscience_'s tweet image. Financial Statement Analysis with Large Language Models (LLMs)

A 54-page PDF:

Mr. L 님이 재게시함

Training LLMs with Reinforcement Learning (RL) isn’t a new idea. So why does it suddenly seem to be working now (o1/DeepSeek)? Here are a few theories and my thoughts on each of them: (1/N)

zhengyaojiang's tweet image. Training LLMs with Reinforcement Learning (RL) isn’t a new idea.
So why does it suddenly seem to be working now (o1/DeepSeek)?

Here are a few theories and my thoughts on each of them: (1/N)

Mr. L 님이 재게시함

Local web research assistant using Ollama LLMs

tom_doerr's tweet image. Local web research assistant using Ollama LLMs

United States 트렌드

Loading...

Something went wrong.


Something went wrong.