thefirehacker's profile picture. Founder-AI Researcher. Building BubblSpace & Timecapsule

FireHacker

@thefirehacker

Founder-AI Researcher. Building BubblSpace & Timecapsule

置頂

Use your favourite AI coding agent to create AI frames. What if you could connect everything—your PDFs, videos, notes, code, and research—into one seamless flow that actually makes sense? AI-Frames: Open Source Knowledge-to-Action Platform:timecapsule.bubblspace.com ✨ Annotate •…


FireHacker 已轉發

On value functions, great answer. Michael Truell: Why aren't value functions popular in RL right now? John Schulman: I'd say they don't seem to help very much in the settings that people are doing RL on right now. So for example, doing RL from human feedback and RL on these…

A conversation with @johnschulman2 on the first year LLMs could have been useful, building research teams, and where RL goes from here. 00:20 - Speedrunning ChatGPT 09:22 - Archetypes of research managers 11:56 - Was OpenAI inspired by Bell Labs? 16:54 - The absence of value…



FireHacker 已轉發

This is the Outstanding Paper Award at ICLR 2025, and this is exactly the kind of research on LLMs we need, not those quasi-psychological studies of the form "we asked the same question to these 3 models and see which one is more racist!" As you might already know, when…

burkov's tweet image. This is the Outstanding Paper Award at ICLR 2025, and this is exactly the kind of research on LLMs we need, not those quasi-psychological studies of the form "we asked the same question to these 3 models and see which one is more racist!"

As you might already know, when…

FireHacker 已轉發

We are thrilled to share significant progress on our ambitious Artificial Intelligence (AI) and Large Language Model (LLM) development program for Sanskrit, which was officially inaugurated on Vijayadasami day this year. The Core Mission and Team A dedicated core group of…

sanskrit1906's tweet image. We are thrilled to share significant progress on our ambitious Artificial Intelligence (AI) and Large Language Model (LLM) development program for Sanskrit, which was officially inaugurated on Vijayadasami day this year.

The Core Mission and Team
A dedicated core group of…

FireHacker 已轉發

VIDEO FOR THE AGES - SACHIN TENDULKAR AND LEO MESSI ON THE SAME STAGE. ❤️


FireHacker 已轉發

Olmo 3.1 is here. We extended our strongest RL run and scaled our instruct recipe to 32B—releasing Olmo 3.1 Think 32B & Olmo 3.1 Instruct 32B, our most capable models yet. 🧵

allen_ai's tweet image. Olmo 3.1 is here. We extended our strongest RL run and scaled our instruct recipe to 32B—releasing Olmo 3.1 Think 32B & Olmo 3.1 Instruct 32B, our most capable models yet. 🧵
allen_ai's tweet image. Olmo 3.1 is here. We extended our strongest RL run and scaled our instruct recipe to 32B—releasing Olmo 3.1 Think 32B & Olmo 3.1 Instruct 32B, our most capable models yet. 🧵

GPT 5.2 Review in one line "Opus 4.5 Zindabad ✊"


FireHacker 已轉發

Today we open source Nomos 1. At just 30B parameters, it scores 87/120 on this year’s Putnam, one of the world’s most prestigious math competitions. This score would rank #2/3988 in 2024 and marks our first step with @hillclimbai towards creating a SOTA AI mathematician.

NousResearch's tweet image. Today we open source Nomos 1. At just 30B parameters, it scores 87/120 on this year’s Putnam, one of the world’s most prestigious math competitions.

This score would rank #2/3988 in 2024 and marks our first step with @hillclimbai towards creating a SOTA AI mathematician.

FireHacker 已轉發

CLOUDFLARE.

himanshustwts's tweet image. CLOUDFLARE.

FireHacker 已轉發

We need more senior researchers camping out at their posters like this. Managed to catch 10 minutes of Alyosha turning @anand_bhattad’s poster into a pop-up mini lecture. Extra spark after he spotted @jathushan. Other folks in the audience: @HaoLi81 @konpatp @GurushaJuneja.


FireHacker 已轉發

One of the coolest papers this NeurIPS so far: FlowFeat.

ana_dodik's tweet image. One of the coolest papers this NeurIPS so far: FlowFeat.

FireHacker 已轉發

Why does NVIDIA spend millions of GPU hours training models just to give it all away? Here's a recording to the talk at NeurIPS for anyone who couldn't make it, wish we recorded the questions afterwards!

I'm at NeurIPS! Giving a talk today on state of OSS & why NVIDIA spends millions of GPU hours training models just to give it all away: datasets, recipes, weights, architecture. 545pm @ Exhibit Hall A,B

NaderLikeLadder's tweet image. I'm at NeurIPS!

Giving a talk today on state of OSS & why NVIDIA spends millions of GPU hours training models just to give it all away: datasets, recipes, weights, architecture.

545pm @ Exhibit Hall A,B


FireHacker 已轉發

This morning at NeurIPS, Rich Sutton reminded us that we need continual learning to reach AGI. This afternoon, Ali Behrouz presented a Google poster paper, Nested Learning, which provides new ideas on the path to continual learning. I recorded the 40 minute talk as it might be…


FireHacker 已轉發

The NeurIPS Test of Time Award recognizes papers that have made lasting contributions to machine learning. For 2025 the award goes to…(buff.ly/W4ohNLA) #NeurIPS2025 #NeurIPSanDiego"

NeurIPSConf's tweet image. The NeurIPS Test of Time Award recognizes papers that have made lasting contributions to machine learning. For 2025 the award goes to…(buff.ly/W4ohNLA)  #NeurIPS2025 #NeurIPSanDiego"

FireHacker 已轉發

Morning traffic at #NeurIPS2025 😅🫠


FireHacker 已轉發

Now to wait for someone to have the best OSS UI for me to run Qwen3 Omni now that vLLM-Omni exists 😍

TheZachMueller's tweet image. Now to wait for someone to have the best OSS UI for me to run Qwen3 Omni now that vLLM-Omni exists 😍

FireHacker 已轉發

Holy sh*t. 🚨Running a Qwen3 fine-tune locally can compromise your entire system. Let's talk about a critical security vulnerability 🧵


As someone who has been crafting AI-Persona applications for the past two years, this felt like a pleasant surprise.I found this gold nugget on the Mexico City track !!! 🔥"First Workshop on LLM Persona Modeling" full day track at NeurIPS2025.Workshop. Workshop…


FireHacker 已轉發

Same for me. I switched 100% to Gemini after 2.5pro. Similar quality as other models (certainly good enough for everyday tasks) but fantastically low latency.

I started using gemini 3 for anything that isn't a hardcore proof. Eg explaining concepts or well known things that don't require any new ideas. Much stabler and faster than 5.1 pro. At least I know it will finish and reply in 5 min and probably faster. With gpt 5.1 pro/thinking…



FireHacker 已轉發

had some discussions that sparked a long thinking session so here is the yield. hope some of you enjoy consuming it.


FireHacker 已轉發

On the streets of San Francisco, I met this really cool researcher going to NeurIPS from @smallest_AI who is working on the intersection between speech models and gravitational lensing! Link to the paper below!


Loading...

Something went wrong.


Something went wrong.