PrahithaM's profile picture. MSCS @UMassAmherst, Community Lead @cohere

Prahitha Movva @COLM2025

@PrahithaM

MSCS @UMassAmherst, Community Lead @cohere

Prahitha Movva @COLM2025 已转帖

Why don’t VLAs generalize as well as their VLM counterparts? One culprit: catastrophic forgetting during fine-tuning. 🧠 We introduce VLM2VLA: a training paradigm that preserves the VLM capabilities while teaching robotic control. vlm2vla.github.io 🧵


Prahitha Movva @COLM2025 已转帖

This paper by Ivan Lee (@ivn1e) & @BergKirkpatrick was great! Best thing I’ve seen at #COLM2025 so far! Readability ≠ Learnability: Rethinking the Role of Simplicity in Training Small Language Models openreview.net/forum?id=AFMGb…


Prahitha Movva @COLM2025 已转帖

Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for…


Prahitha Movva @COLM2025 已转帖

Headed to COLM this week! I’ll be presenting at the NLP4Democracy Workshop and would love to connect over coffee to chat about societal impact of AI, alignment, or AI policy. Also, currently on the job market, always up for chats about research or opportunities! #COLM2025


Prahitha Movva @COLM2025 已转帖

At @ChandarLab, we are happy to announce the third edition of our assistance program to provide feedback for members of communities underrepresented in AI who want to apply to high-profile graduate programs. Want feedback? Details: chandar-lab.github.io/grad_app/. Deadline: Nov 01! cc:…

apsarathchandar's tweet image. At @ChandarLab, we are happy to announce the third edition of our assistance program to provide feedback for members of communities underrepresented in AI who want to apply to high-profile graduate programs.  Want feedback? Details: chandar-lab.github.io/grad_app/. Deadline: Nov 01!

cc:…

Prahitha Movva @COLM2025 已转帖

Over the past few months, I’ve heard the same complaint from nearly every collaborator working on computational cogsci + behavioral and mechanistic interpretability: “Open-source VLMs are a pain to run, let alone analyze.” We finally decided to do something about it (thanks…

ziqiao_ma's tweet image. Over the past few months, I’ve heard the same complaint from nearly every collaborator working on computational cogsci + behavioral and mechanistic interpretability: 

“Open-source VLMs are a pain to run, let alone analyze.”

We finally decided to do something about it (thanks…

Prahitha Movva @COLM2025 已转帖

It is PhD application season again 🍂 For those looking to do a PhD in AI, these are some useful resources 🤖: 1. Examples of statements of purpose (SOPs) for computer science PhD programs: cs-sop.org [1/4]


Prahitha Movva @COLM2025 已转帖

Smol win, just realized the work I published last year got its first citation :) I really want to contribute to the world of Mechanistic Interpretability, and the thought of being cited along with the giants of the field, for a field of work I am so passionate about, made my day!…

akankshanc's tweet image. Smol win, just realized the work I published last year got its first citation :)
I really want to contribute to the world of Mechanistic Interpretability, and the thought of being cited along with the giants of the field, for a field of work I am so passionate about, made my day!…

Prahitha Movva @COLM2025 已转帖

It’s PhD / grad school application season again🌱 I’ve been getting reach out about SoPs, research fit, “should I do PhD at all??” Here's the dump of my honest thoughts from my application experience - what worked for me, and how I decided if a PhD was right for me. Wrote it…


Prahitha Movva @COLM2025 已转帖

@ folks seeking PhD & post-doc opportunities! My friends @sandrabae_sb and @TakanoriFW are starting a new lab! both are such talented, thoughtful researchers and I know they'll be generous and supportive advisors! students will be so lucky to get the chance to work with them!

xjasminelu's tweet image. @ folks seeking PhD & post-doc opportunities! My friends @sandrabae_sb and @TakanoriFW are starting a new lab! both are such talented, thoughtful researchers and I know they'll be generous and supportive advisors! students will be so lucky to get the chance to work with them!

Prahitha Movva @COLM2025 已转帖

Reward models do not have the capacity to fully capture human preferences. If they can't represent human preferences, how can we hope to use them to align a language model? In our #COLM2025 "Off-Policy Corrected Reward Modeling for RLHF", we investigate this issue 🧵

johannesack's tweet image. Reward models do not have the capacity to fully capture human preferences.
If they can't represent human preferences, how can we hope to use them to align a language model? 

In our #COLM2025 "Off-Policy Corrected Reward Modeling for RLHF", we investigate this issue 🧵

Prahitha Movva @COLM2025 已转帖

Who is going to be at #COLM2025? I want to draw your attention to a COLM paper by my student @sheridan_feucht that has totally changed the way I think and teach about LLM representations. The work is worth knowing. And you meet Sheridan at COLM, Oct 7!

davidbau's tweet image. Who is going to be at #COLM2025?

I want to draw your attention to a COLM paper by my student @sheridan_feucht that has totally changed the way I think and teach about LLM representations. The work is worth knowing.

And you meet Sheridan at COLM, Oct 7!

[📄] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.

sheridan_feucht's tweet image. [📄] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.


Prahitha Movva @COLM2025 已转帖

🔥New preprint: Soft Tokens, Hard Truths Introduces the first scalable continuous-token RL method for LLMs - no reference CoTs needed; scales to hundreds of thought tokens. Best to train soft, infer hard! Pass@1 parity ⚖️, Pass@32 gains 📈& better robustness 🛡️ vs. hard CoT 1/🧵

NatashaEve4's tweet image. 🔥New preprint: Soft Tokens, Hard Truths
Introduces the first scalable continuous-token RL method for LLMs - no reference CoTs needed; scales to hundreds of thought tokens. Best to train soft, infer hard! Pass@1 parity ⚖️, Pass@32 gains 📈& better robustness 🛡️ vs. hard CoT
1/🧵
NatashaEve4's tweet image. 🔥New preprint: Soft Tokens, Hard Truths
Introduces the first scalable continuous-token RL method for LLMs - no reference CoTs needed; scales to hundreds of thought tokens. Best to train soft, infer hard! Pass@1 parity ⚖️, Pass@32 gains 📈& better robustness 🛡️ vs. hard CoT
1/🧵

Prahitha Movva @COLM2025 已转帖

Are there conceptual directions in VLMs that transcend modality? Check out our COLM spotlight🔦 paper! We analyze how linear concepts interact with multimodality in VLM embeddings using SAEs with @Huangyu58589918, @napoolar, @ShamKakade6 and Stephanie Gil arxiv.org/abs/2504.11695

isabelpapad's tweet image. Are there conceptual directions in VLMs that transcend modality? Check out our COLM spotlight🔦 paper! We analyze how linear concepts interact with multimodality in VLM embeddings using SAEs

with @Huangyu58589918, @napoolar, @ShamKakade6 and Stephanie Gil
arxiv.org/abs/2504.11695

Prahitha Movva @COLM2025 已转帖

We’ve received A LOT OF submissions this year 🤯🤯 and are excited to see so much interest! To ensure high-quality review, we are looking for more dedicated reviewers. If you'd like to help, please sign up here docs.google.com/forms/d/e/1FAI…


Prahitha Movva @COLM2025 已转帖

✨ Internship Opportunity @ Google Research ✨ We are seeking a self-motivated student researcher to join our team at Google Research starting around January 2026. 🚀 In this role, you will contribute to research projects advancing agentic LLMs through tool use and RL, with the…


Prahitha Movva @COLM2025 已转帖

On my way to #UIST2025, kicking off strong with a 16 hour+ flight due to a refueling stop. ✈️ In Seoul until 28th then Busan. HCI friends, lmk if you’d like to meet up! We are also hiring a PhD student for projects around human-AI grounding —will post this again closer to conf.


Prahitha Movva @COLM2025 已转帖

In today's modern AI world I sometimes wonder if "research scientist or engineer" is the right dichotomy anymore. Im not convinced this is the right framework anymore. I think it should be a slider of a single factor called "creativity" and that's it. It's also not exactly a…


Prahitha Movva @COLM2025 已转帖

I'm looking for an informal PhD supervisor in LLMs/post-training — any recommendations? My supervisor is leaving academia & the rest of the dep't doesn't work on LLMs, so I'm hoping to find someone external to collaborate with More info 👇, RTs appreciated! 🙏


Prahitha Movva @COLM2025 已转帖

Our ML Industry program is excited to host @psssnikhil1, Senior ML Engineer at Adobe for a session on "Crafting a Successful AI Career & Transitioning Research into Scalable AI Products." Thanks to @__Vaibhavi, @arya_suneesh and @PrahithaM for organizing this event ✨ Learn…

Cohere_Labs's tweet image. Our ML Industry program is excited to host @psssnikhil1, Senior ML Engineer at Adobe for a session on "Crafting a Successful AI Career & Transitioning Research into Scalable AI Products."

Thanks to @__Vaibhavi, @arya_suneesh and @PrahithaM  for organizing this event ✨

Learn…

Loading...

Something went wrong.


Something went wrong.