juand_r_nlp's profile picture. CS PhD student at UT Austin in #NLP
Interested in language, reasoning, semantics and cognitive science. 

You can also find me over at the other site 🦋

Juan Diego Rodríguez (he/him)

@juand_r_nlp

CS PhD student at UT Austin in #NLP Interested in language, reasoning, semantics and cognitive science. You can also find me over at the other site 🦋

Juan Diego Rodríguez (he/him) podał dalej

COLM Keynote: Nicholas Carlini Are LLMs worth it? youtube.com/watch?v=PngHcm…

COLM_conf's tweet card. Nicholas Carlini - Are LLMs worth it?

youtube.com

YouTube

Nicholas Carlini - Are LLMs worth it?


Juan Diego Rodríguez (he/him) podał dalej

I’ll be in Boston attending BUCLD this week — I won’t be presenting but I’ll be cheering on @najoungkim who will present at the prestigious SLD symposium about the awesome work by her group, including our work on LMs as hypotheses generators for language acquisition! 🤠👻

kanishkamisra's tweet image. I’ll be in Boston attending BUCLD this week — I won’t be presenting but I’ll be cheering on @najoungkim who will present at the prestigious SLD symposium about the awesome work by her group, including our work on LMs as hypotheses generators for language acquisition! 

🤠👻

Juan Diego Rodríguez (he/him) podał dalej

Dear “15-18 yo founder”s sending me DMs, don’t. Go and hug your parents, fall in love, eat chocolate cereal for breakfast, read poetry. Nobody will give you back these years. And sure, do your homework and learn math and code if that feels fun. But stop building SaaS and…


Juan Diego Rodríguez (he/him) podał dalej

“For a glorious decade in the 2010s, we all worked on better neural network architectures, but after that we just worked on scaling transformers and making them more efficient”

We already called it in 2020!

chrmanning's tweet image. We already called it in 2020!


Juan Diego Rodríguez (he/him) podał dalej
m2saxon's tweet image.

The Computer Science section of @arxiv is now requiring prior peer review for Literature Surveys and Position Papers. Details in a new blog post



Juan Diego Rodríguez (he/him) podał dalej

236 direct/indirect PhD students!! Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/ Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!

prakashkagitha's tweet image. 236 direct/indirect PhD students!!

Based on OpenReview data, an interactive webpage: prakashkagitha.github.io/manningphdtree/

Your course, NLP with Deep learning - Winter 2017, on YouTube, was my introduction to building deep learning models. This is the least I could do to say Thank You!!

Thanks! 😊 But it’d be really good to generate an updated version of those graphs!



Juan Diego Rodríguez (he/him) podał dalej

I'm super excited to update that "Lost in Automatic Translation" is now available as an audiobook! 🔊📖 It's currently on Audible: audible.ca/pd/B0FXY8VQX5 Stay tuned (lostinautomatictranslation.com) for more retailers, including Amazon, iTunes, etc., and public libraries! 📚

VeredShwartz's tweet image. I'm super excited to update that "Lost in Automatic Translation" is now available as an audiobook! 🔊📖

It's currently on Audible: 
audible.ca/pd/B0FXY8VQX5

Stay tuned (lostinautomatictranslation.com) for more retailers, including Amazon, iTunes, etc., and public libraries! 📚

Juan Diego Rodríguez (he/him) podał dalej

A thread on the equilibria of pendulums and their connection to topology. 1/n


Juan Diego Rodríguez (he/him) podał dalej

We’re drowning in language models — there are over 2 mil. of them on Huggingface! Can we use some of them to understand which computational ingredients — architecture, scale, post-training, etc. – help us build models that align with human representations? Read on to find out 🧵

ZachStuddiford's tweet image. We’re drowning in language models — there are over 2 mil. of them on Huggingface! Can we use some of them to understand which computational ingredients — architecture, scale, post-training, etc. – help us build models that align with human representations? Read on to find out 🧵

Juan Diego Rodríguez (he/him) podał dalej

one thing that really became clear to me (which admittedly makes me publish much less) is that, especially as academics, "beating the state of the art" is a crap target to aim for. the objective should be to replace the state of the art. (of course, this is unfortunately super…


Juan Diego Rodríguez (he/him) podał dalej

𝑵𝒆𝒘 𝒃𝒍𝒐𝒈𝒑𝒐𝒔𝒕! In which I give some brief reflections on #COLM2025 and give a rundown of a few great papers I checked out!

m2saxon's tweet image. 𝑵𝒆𝒘 𝒃𝒍𝒐𝒈𝒑𝒐𝒔𝒕!

In which I give some brief reflections on #COLM2025 and give a rundown of a few great papers I checked out!

Juan Diego Rodríguez (he/him) podał dalej

I will be giving a short talk on this work at the COLM Interplay workshop on Friday (also to appear at EMNLP)! Will be in Montreal all week and excited to chat about LM interpretability + it’s interaction with human cognition and ling theory.

A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advantage of recent advances in mechanistic interpretability to test this hypothesis in Language Models. New work with @kmahowald and @ChrisGPotts! 🧵👇

SashaBoguraev's tweet image. A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advantage of recent advances in mechanistic interpretability to test this hypothesis in Language Models.

New work with @kmahowald and @ChrisGPotts!

🧵👇


Juan Diego Rodríguez (he/him) podał dalej

Happy to announce the first workshop on Pragmatic Reasoning in Language Models — PragLM @ COLM 2025! 🧠🎉 How do LLMs engage in pragmatic reasoning, and what core pragmatic capacities remain beyond their reach? 🌐 sites.google.com/berkeley.edu/p… 📅 Submit by June 23rd


Juan Diego Rodríguez (he/him) podał dalej

Interested in language models, brains, and concepts? Check out our COLM 2025 🔦 Spotlight paper! (And if you’re at COLM, come hear about it on Tuesday – sessions Spotlight 2 & Poster 2)!

maria_ryskina's tweet image. Interested in language models, brains, and concepts? Check out our COLM 2025 🔦 Spotlight paper!

(And if you’re at COLM, come hear about it on Tuesday – sessions Spotlight 2 & Poster 2)!

Juan Diego Rodríguez (he/him) podał dalej

Our paper "ChartMuseum 🖼️" is now accepted to #NeurIPS2025 Datasets and Benchmarks Track! Even the latest models, such as GPT-5 and Gemini-2.5-Pro, still cannot do well on challenging 📉chart understanding questions , especially on those that involve visual reasoning 👀!

LiyanTang4's tweet image. Our paper "ChartMuseum 🖼️" is now accepted to #NeurIPS2025 Datasets and Benchmarks Track!

Even the latest models, such as GPT-5 and Gemini-2.5-Pro, still cannot do well on challenging 📉chart understanding questions , especially on those that involve visual reasoning 👀!

Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts! ✍🏻Entirely human-written questions by 13 CS researchers 👀Emphasis on visual reasoning – hard to be verbalized via text CoTs 📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B

LiyanTang4's tweet image. Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts!

✍🏻Entirely human-written questions by 13 CS researchers
👀Emphasis on visual reasoning – hard to be verbalized via text CoTs
📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B
LiyanTang4's tweet image. Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts!

✍🏻Entirely human-written questions by 13 CS researchers
👀Emphasis on visual reasoning – hard to be verbalized via text CoTs
📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B


Juan Diego Rodríguez (he/him) podał dalej

Accepted at #NeurIPS2025 -- super proud of Yulu and Dheeraj for leading this! Be on the lookout for more "nuanced yes/no" work from them in the future 👀

Does vision training change how language is represented and used in meaningful ways?🤔 The answer is a nuanced yes! Comparing VLM-LM minimal pairs, we find that while the taxonomic organization of the lexicon is similar, VLMs are better at _deploying_ this knowledge. [1/9]

yulu_qin's tweet image. Does vision training change how language is represented and used in meaningful ways?🤔 The answer is a nuanced yes! Comparing VLM-LM minimal pairs, we find that while the taxonomic organization of the lexicon is similar, VLMs are better at _deploying_ this knowledge. [1/9]


Juan Diego Rodríguez (he/him) podał dalej

You shall know a fascist asshole by the company he keeps... And also what he says.

Elon Musk, "There's got to be a change of government in Britain" "We don't have another four years or whenever your next election is, it's too long, something has got to be done" "There has got to be a dissolution of parliament and a new vote held" "You got to appeal to the…



Juan Diego Rodríguez (he/him) podał dalej

Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts! ✍🏻Entirely human-written questions by 13 CS researchers 👀Emphasis on visual reasoning – hard to be verbalized via text CoTs 📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B

LiyanTang4's tweet image. Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts!

✍🏻Entirely human-written questions by 13 CS researchers
👀Emphasis on visual reasoning – hard to be verbalized via text CoTs
📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B
LiyanTang4's tweet image. Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts!

✍🏻Entirely human-written questions by 13 CS researchers
👀Emphasis on visual reasoning – hard to be verbalized via text CoTs
📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B

Juan Diego Rodríguez (he/him) podał dalej

Solving complex problems with CoT requires combining different skills. We can do this by: 🧩Modify the CoT data format to be “composable” with other skills 🔥Train models on each skill 📌Combine those models Lead to better 0-shot reasoning on tasks involving skill composition!

fangcong_y10593's tweet image. Solving complex problems with CoT requires combining different skills.

We can do this by:
🧩Modify the CoT data format to be “composable” with other skills
🔥Train models on each skill
📌Combine those models

Lead to better 0-shot reasoning on tasks involving skill composition!
fangcong_y10593's tweet image. Solving complex problems with CoT requires combining different skills.

We can do this by:
🧩Modify the CoT data format to be “composable” with other skills
🔥Train models on each skill
📌Combine those models

Lead to better 0-shot reasoning on tasks involving skill composition!

Juan Diego Rodríguez (he/him) podał dalej

Super thrilled that @kanishkamisra is going to join @UT_Linguistics as our newest computational linguistics faculty member -- looking forward to doing great research together! 🧑‍🎓Students: Kanishka is a GREAT mentor -- apply to be his PhD student in the upcoming cycle!!

News🗞️ I will return to UT Austin as an Assistant Professor of Linguistics this fall, and join its vibrant community of Computational Linguists, NLPers, and Cognitive Scientists!🤘 Excited to develop ideas about linguistic and conceptual generalization! Recruitment details soon

kanishkamisra's tweet image. News🗞️

I will return to UT Austin as an Assistant Professor of Linguistics this fall, and join its vibrant community of Computational Linguists, NLPers, and Cognitive Scientists!🤘

Excited to develop ideas about linguistic and conceptual generalization! Recruitment details soon


Loading...

Something went wrong.


Something went wrong.