benjiewang_cs's profile picture. Postdoc at UCLA StarAI Lab @UCLAComSci

Benjie Wang

@benjiewang_cs

Postdoc at UCLA StarAI Lab @UCLAComSci

Benjie Wang 已轉發

“That’s one small [MASK] for [MASK], a giant [MASK] for mankind.” – [MASK] Armstrong Can autoregressive models predict the next [MASK]? It turns out yes, and quite easily… Introducing MARIA (Masked and Autoregressive Infilling Architecture) arxiv.org/abs/2502.06901


Benjie Wang 已轉發

So excited to present Ctrl-G **Adaptable Logical Control for Large Language Models** TODAY at #NeurIPS2024 West Ballroom 4:30 - 7:30 pm. Ctrl-G is THE solution to LLM fill-in-the-middle generation, numerical planning and structured output. Stop by to discuss more!

Proposing Ctrl-G, a neurosymbolic framework that enables arbitrary LLMs to follow logical constraints (length control, infilling …) with 100% guarantees. Ctrl-G beats GPT4 on the task of text editing by >30% higher satisfaction rate in human eval. arxiv.org/abs/2406.13892



Benjie Wang 已轉發

📢 I’m recruiting PhD students @CS_UVA for Fall 2025! 🎯 Neurosymbolic AI, probabilistic ML, trustworthiness, AI for science. See my website for more details: zzeng.me 📬 If you're interested, apply and mention my name in your application: engineering.virginia.edu/department/com…


Benjie Wang 已轉發

🚨 Exciting Opportunity! 🚨 I’m looking for PhD students to join my team @ImperialEEE and @ImperialX_AI! 🌍🔍 Research Topic: Neuro-symbolic AI with a focus on making AI safer. 💡🤖 Full scholarships available! 🎓💰 Interested? Email me at: [email protected] [1/2]


Excited to share our work on LLM tokenization, led by the awesome @renatogeh. We find significant boosts in downstream performance, by probabilistically interpreting the space of tokenizations of a text. A bit of probabilistic reasoning goes a long way!

Where is the signal in LLM tokenization space? Does it only come from the canonical (default) tokenization? The answer is no! By looking at other ways to tokenize the same text, we get a consistent boost to LLM performance! arxiv.org/abs/2408.08541 1/5

renatogeh's tweet image. Where is the signal in LLM tokenization space?

Does it only come from the canonical (default) tokenization?

The answer is no! By looking at other ways to tokenize the same text, we get a consistent boost to LLM performance!

arxiv.org/abs/2408.08541

1/5
renatogeh's tweet image. Where is the signal in LLM tokenization space?

Does it only come from the canonical (default) tokenization?

The answer is no! By looking at other ways to tokenize the same text, we get a consistent boost to LLM performance!

arxiv.org/abs/2408.08541

1/5
renatogeh's tweet image. Where is the signal in LLM tokenization space?

Does it only come from the canonical (default) tokenization?

The answer is no! By looking at other ways to tokenize the same text, we get a consistent boost to LLM performance!

arxiv.org/abs/2408.08541

1/5


Super cool work on discretizing probability distributions with *exponential* gains in succinctness! Recommended reading for probabilistic inference folks

Are you looking for an inference algorithm that supports your discrete-continuous probabilistic program? Look no further! We have developed a new probabilistic programming language (PPL) called HyBit that provides scalable support for discrete-continuous probabilistic programs.



Benjie Wang 已轉發

We will organize a "Causality for LLMs" Tutorial #NeurIPS2024 @NeurIPSConf. Happy to contribute to our community an intro of meaningful topics in Causal LLMs. And super excited for our panel w/ Yoshua, @bschoelkopf @LukeZettlemoyer @_jasonwei @Swarooprm7 @tom4everitt. Stay tuned!

ZhijingJin's tweet image. We will organize a "Causality for LLMs" Tutorial #NeurIPS2024 @NeurIPSConf. Happy to contribute to our community an intro of meaningful topics in Causal LLMs. And super excited for our panel w/ Yoshua, @bschoelkopf @LukeZettlemoyer @_jasonwei @Swarooprm7 @tom4everitt. Stay tuned!

Loading...

Something went wrong.


Something went wrong.