stanfordnlp's profile picture. Computational Linguists—Natural Language—Machine Learning @chrmanning @jurafsky @percyliang @ChrisGPotts @tatsu_hashimoto @MonicaSLam @Diyi_Yang @StanfordAILab

Stanford NLP Group

@stanfordnlp

Computational Linguists—Natural Language—Machine Learning @chrmanning @jurafsky @percyliang @ChrisGPotts @tatsu_hashimoto @MonicaSLam @Diyi_Yang @StanfordAILab

Stanford NLP Group 님이 재게시함

Highly recommend the Tinker Research and Teaching Grants: thinkingmachines.ai/blog/tinker-re… We found it very helpful when updating Stanford CS 329X: Human-Centered LLMs this year - many students requested to keep using it with creative course project ideas after the assignment.

EchoShao8899's tweet image. Highly recommend the Tinker Research and Teaching Grants: thinkingmachines.ai/blog/tinker-re…

We found it very helpful when updating Stanford CS 329X: Human-Centered LLMs this year - many students requested to keep using it with creative course project ideas after the assignment.

Today we’re announcing research and teaching grants for Tinker: credits for scholars and students to fine-tune and experiment with open-weight LLMs. Read more and apply at: thinkingmachines.ai/blog/tinker-re…



Stanford NLP Group 님이 재게시함

The innovation ecosystem behind today’s dramatic AI wave was developed through open science and open development. With corporate AI labs turning secretive, universities and nonprofits must reclaim innovating with open AI research for the public good. hai.stanford.edu/news/universit…


Stanford NLP Group 님이 재게시함

Marin project is insanely cool! and they get stuff done

⛵Marin 32B Base (mantis) is done training! It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base. Ranking across 19 benchmarks:

percyliang's tweet image. ⛵Marin 32B Base (mantis) is done training!  It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base.  Ranking across 19 benchmarks:


Stanford NLP Group 님이 재게시함

Yesterday I asked what makes you happy. 226 voted--61% said family & friends, like those AI researchers at the panel. Not AI, just the people around us. This resonates with Harvard's 87-year study on happiness. Their conclusion: happiness is not about money/fame/success, it's…

shi_weiyan's tweet image. Yesterday I asked what makes you happy. 226 voted--61% said family & friends, like those AI researchers at the panel. Not AI, just the people around us.

This resonates with Harvard's 87-year study on happiness. Their conclusion: happiness is not about money/fame/success, it's…

I heard the best question at an AI panel recently: "what makes you happy these days?" Top AI researchers answered: my partner, my family, farmers market, good food. Nothing about AI—everything about the humans around us. In an industry burning out, maybe that's the answer?



Stanford NLP Group 님이 재게시함

@karpathy observed LLMs are "silently collapsed...only know 3 jokes". We prove this is mathematically inevitable due to RLHF + human psychology. But these capabilities aren't lost, just hidden – and easily restored. This means AI benchmarks are measuring training artifacts.🧵

shi_weiyan's tweet image. @karpathy observed LLMs are "silently collapsed...only know 3 jokes".

We prove this is mathematically inevitable due to RLHF + human psychology.

But these capabilities aren't lost, just hidden – and easily restored.

This means AI benchmarks are measuring training artifacts.🧵

Stanford NLP Group 님이 재게시함

🥳I have just added Marin gptq and awq quantization support to GPT-QModel with the first release of a Marin W4A16 gptq quant on HF. dl link in comments. 👇 @percyliang @dlwh

⛵Marin 32B Base (mantis) is done training! It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base. Ranking across 19 benchmarks:

percyliang's tweet image. ⛵Marin 32B Base (mantis) is done training!  It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base.  Ranking across 19 benchmarks:


Stanford NLP Group 님이 재게시함

⛵Marin 32B Base (mantis) is done training! It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base. Ranking across 19 benchmarks:

percyliang's tweet image. ⛵Marin 32B Base (mantis) is done training!  It is the best open-source base model (beating OLMo 2 32B Base) and it’s even close to the best comparably-sized open-weight base models, Gemma 3 27B PT and Qwen 2.5 32B Base.  Ranking across 19 benchmarks:

Stanford NLP Group 님이 재게시함

1/4 Following up on our launch of Tongyi DeepResearch: We're now releasing the full technical report! Dive deep into the technology and insights behind our 30B (A3B) open-source web agent that achieves SOTA performance: 32.9 on Humanity's Last Exam, 43.4 on BrowseComp, and 46.7…

Ali_TongyiLab's tweet image. 1/4 Following up on our launch of Tongyi DeepResearch: We're now releasing the full technical report! Dive deep into the technology and insights behind our 30B (A3B) open-source web agent that achieves SOTA performance: 32.9 on Humanity's Last Exam, 43.4 on BrowseComp, and 46.7…

Stanford NLP Group 님이 재게시함

Someone should do an analysis of the cumulative impact of Chris Manning - Research he co-produced - Students he advised who went on to be top faculty and industry leaders - Untold distributed impact from textbooks, open-source software, and online lectures Absolutely incredible

236 direct/indirect PhD students!! Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/ Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!

prakashkagitha's tweet image. 236 direct/indirect PhD students!!

Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/

Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!


Stanford NLP Group 님이 재게시함

Marin, which @percyliang, @dlwh, and many others are building is *fully open* (not just open weights) and has been used to build high-quality, competitive models. Come to the talk next week at Ray Summit 😀

robertnishihara's tweet image. Marin, which @percyliang, @dlwh, and many others are building is *fully open* (not just open weights) and has been used to build high-quality, competitive models.

Come to the talk next week at Ray Summit 😀

Open AI means AI that is open.



Stanford NLP Group 님이 재게시함

Open AI means AI that is open.


Stanford NLP Group 님이 재게시함

What I found most surprising from this graph is that @percyliang has somehow already eclipsed Dan Klein in number of graduated PhD students and is close to catching @chrmanning Different styles in how they have built their groups, in different eras, but nonetheless surprising

236 direct/indirect PhD students!! Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/ Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!

prakashkagitha's tweet image. 236 direct/indirect PhD students!!

Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/

Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!


Stanford NLP Group 님이 재게시함

我原定的计划是CS229->CS224n->CS336(期间的论文阅读量能应付两位导师就好),学完这些课之后再all in多智能体论文

同样是机器学习 感觉差别好大 我们学校的这门课就是把西瓜书转化成ppt然后念 遂只听了第一节课 而CS229看了7个lecture感觉就像一堆概率论夹杂着一点微积分+线代 对于概率论“0基础”的我来说每个lecture都能看一下午 但是不管是哪一个 和我们组多智能体的研究方向都相差甚远 迷茫



Stanford NLP Group 님이 재게시함

Excited to talk about Augmenting Humans with LLMs 💪 at 2pm today at The University of Michigan Symposium on Human-Centered AI 🤩 What a beautiful campus! 📍

Diyi_Yang's tweet image. Excited to talk about Augmenting Humans with LLMs 💪  at 2pm today at The University of Michigan Symposium on Human-Centered AI 🤩

What a beautiful campus! 📍

Stanford NLP Group 님이 재게시함

We recorded a bunch of people actually working on their computers (!) and then compared agent performance to actual human workflows. Awesome paper led by @ZhiruoW :)

Agents are joining us at work -- coding, writing, design. But how do they actually work, especially compared to humans? Their workflows tell a different story: They code everything, slow down human flows, and deliver low-quality work fast. Yet when teamed with humans, they shine…



Stanford NLP Group 님이 재게시함

Thanks! I want to stick with the overall assessment on the slide, though, especially the juxtaposition of the final part of the Lewis quotation with my concern about it on the right. I suspect my concern extends to your claim about what we can/should keep separate!

ChrisGPotts's tweet image. Thanks! I want to stick with the overall assessment on the slide, though, especially the juxtaposition of the final part of the Lewis quotation with my concern about it on the right. I suspect my concern extends to your claim about what we can/should keep separate!

Stanford NLP Group 님이 재게시함

@ZhiruoW 's research compares AI agents vs humans across real work tasks (data analysis, engineering, design, writing). Key findings: 👉Agents are 88% faster & 90-96% cheaper 👉BUT produce lower quality work, often fabricate data to mask limitations 👉Agents code everything,…

Diyi_Yang's tweet image. @ZhiruoW 's  research compares AI agents vs humans across real work tasks (data analysis, engineering, design, writing). Key findings:

👉Agents are 88% faster & 90-96% cheaper
👉BUT produce lower quality work, often fabricate data to mask limitations
👉Agents code everything,…

Agents are joining us at work -- coding, writing, design. But how do they actually work, especially compared to humans? Their workflows tell a different story: They code everything, slow down human flows, and deliver low-quality work fast. Yet when teamed with humans, they shine…



Stanford NLP Group 님이 재게시함

CS 276 from @stanfordnlp is where I met my co-founder at Turing! (@krishnanvijay) Such a fun class taught by @chrmanning & @WittedNote which sparked a lifelong passion in AI. (Along with CS 229 by @AndrewYNg)

Today, we’re overjoyed to have a 25th Anniversary Reunion of @stanfordnlp. So happy to see so many of our former students back at @Stanford. And thanks to @StanfordHAI for the venue!

stanfordnlp's tweet image. Today, we’re overjoyed to have a 25th Anniversary Reunion of @stanfordnlp. 

So happy to see so many of our former students back at @Stanford. 

And thanks to @StanfordHAI for the venue!


Stanford NLP Group 님이 재게시함

Analyzing agent trajectories has become a new canonical way for evaluation. Yet most trajectories operate at the action level - dense but semantically shallow. We define 𝙬𝙤𝙧𝙠𝙛𝙡𝙤𝙬 as a higher-level abstraction and release a toolkit to induce it from raw actions for both…

EchoShao8899's tweet image. Analyzing agent trajectories has become a new canonical way for evaluation. Yet most trajectories operate at the action level - dense but semantically shallow.
We define 𝙬𝙤𝙧𝙠𝙛𝙡𝙤𝙬 as a higher-level abstraction and release a toolkit to induce it from raw actions for both…

Agents are joining us at work -- coding, writing, design. But how do they actually work, especially compared to humans? Their workflows tell a different story: They code everything, slow down human flows, and deliver low-quality work fast. Yet when teamed with humans, they shine…



Loading...

Something went wrong.


Something went wrong.