ZhiyuanCS's profile picture. PhD student in @NUSingapore Visiting Researcher in @MIT

Zhiyuan

@ZhiyuanCS

PhD student in @NUSingapore Visiting Researcher in @MIT

Zhiyuan hat repostet

🚀Introducing Lumine, a generalist AI agent trained within Genshin Impact that can perceive, reason, and act in real time, completing hours-long missions and following diverse instructions within complex 3D open-world environments.🎮 Website: lumine-ai.org 1/6


How to determine which idea is most promising to scale up? Feedback from the chat with Sora 2 researcher: Even in big tech, you must prove a method's worth scaling up. Key hint? Under fixed compute and targeted perspectives (e.g., deep reasoning in LLMs or physical…


Zhiyuan hat repostet

🚀Introducing GTA1 – our new GUI Agent that leads the OSWorld leaderboard with a 45.2% success rate, outperforming OpenAI's CUA! GTA1 improves two core components of GUI agents: Planning and Grounding. 🧠 Planning: A generic test-time scaling strategy that concurrently samples…

LiJunnan0409's tweet image. 🚀Introducing GTA1 – our new GUI Agent that leads the OSWorld leaderboard with a 45.2% success rate, outperforming OpenAI's CUA!

GTA1 improves two core components of GUI agents: Planning and Grounding.

🧠 Planning: A generic test-time scaling strategy that concurrently samples…

Zhiyuan hat repostet

Customizing Your LLMs in seconds using prompts🥳! Excited to share our latest work with @HPCAILab, @VITAGroupUT, @k_schuerholt, @YangYou1991, @mmbronstein, @damianborth : Drag-and-Drop LLMs(DnD). 2 features: tuning-free, comparable or even better than full-shot tuning.(🧵1/8)


🚨🚨Reviewed around 20 papers for @ACMMM—but our own reviews were hidden & forced on us without expertise match. Time to rethink AI community peer review. 🤔 Our author team were assigned nearly 20 papers with no regard for our areas of expertise, received only a single round of…


I can’t believe this jaw‑dropping comic was generated by GPT just by feeding it our paper directly🤯! It perfectly illustrates how meta‑ability training makes LRMs think better.

ZhiyuanCS's tweet image. I can’t believe this jaw‑dropping comic was generated by GPT just by feeding it our paper directly🤯! 

It perfectly illustrates how meta‑ability training makes LRMs think better.

🚀 Beyond “aha”: toward Meta‑Abilities Alignment! Zero human annotation enables LRMs masters strong reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬. Meta‑ability alignment lifts the ceiling of further domain‑RL—7B → 32B…

ZhiyuanCS's tweet image. 🚀 Beyond “aha”: toward Meta‑Abilities Alignment!
Zero human annotation enables LRMs masters strong reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬.

Meta‑ability alignment lifts the ceiling of further domain‑RL—7B → 32B…
ZhiyuanCS's tweet image. 🚀 Beyond “aha”: toward Meta‑Abilities Alignment!
Zero human annotation enables LRMs masters strong reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬.

Meta‑ability alignment lifts the ceiling of further domain‑RL—7B → 32B…

🚀 Beyond 'aha': toward Meta‑Abilities Alignment! By self‑synthesizes training tasks & self‑verifies rewards with zero human labels, LLM systematically masters core reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬. Meta‑ability…

ZhiyuanCS's tweet image. 🚀 Beyond 'aha': toward Meta‑Abilities Alignment!
By self‑synthesizes training tasks & self‑verifies rewards with zero human labels, LLM systematically masters core reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬.
Meta‑ability…
ZhiyuanCS's tweet image. 🚀 Beyond 'aha': toward Meta‑Abilities Alignment!
By self‑synthesizes training tasks & self‑verifies rewards with zero human labels, LLM systematically masters core reasoning abilities rather than aha emerging and generalize across math ⚙️, code 💻, science 🔬.
Meta‑ability…

Although the ICLR main conference is coming to an end, we are excited to invite you to the Reasoning and Planning for LLMs Workshop, which will be held all day on Monday, April 28. We are honored to host an outstanding lineup of keynote speakers and panelists from Meta, OpenAI,…

ZhiyuanCS's tweet image. Although the ICLR main conference is coming to an end, we are excited to invite you to the Reasoning and Planning for LLMs Workshop, which will be held all day on Monday, April 28.

We are honored to host an outstanding lineup of keynote speakers and panelists from Meta, OpenAI,…

Zhiyuan hat repostet

Welcome to use JudgeLRM! Compare any Hugging Face language models by asking your own questions, and explore JudgeLRM’s reasoning and detailed comparisons! Demo: huggingface.co/spaces/nuojohn… Paper: huggingface.co/papers/2504.00… Model: huggingface.co/nuojohnchen/Ju… Code: github.com/NuoJohnChen/Ju… We…

NuoJohnChen's tweet image. Welcome to use JudgeLRM! Compare any Hugging Face language models by asking your own questions, and explore JudgeLRM’s reasoning and detailed comparisons!
Demo: huggingface.co/spaces/nuojohn…
Paper: huggingface.co/papers/2504.00…
Model: huggingface.co/nuojohnchen/Ju…
Code: github.com/NuoJohnChen/Ju…

We…
NuoJohnChen's tweet image. Welcome to use JudgeLRM! Compare any Hugging Face language models by asking your own questions, and explore JudgeLRM’s reasoning and detailed comparisons!
Demo: huggingface.co/spaces/nuojohn…
Paper: huggingface.co/papers/2504.00…
Model: huggingface.co/nuojohnchen/Ju…
Code: github.com/NuoJohnChen/Ju…

We…

🚀 Exciting news! The ICLR 2025 LLM Reasoning & Planning Workshop is offering several Student Registration Grants to support early-career researchers 💡 Free ICLR registration for in-person full-time students! Apply by March 2, 2025. More info: …shop-llm-reasoning-planning.github.io Submit…


🚀 Call for Reviewers! 🚀 Our Workshop on Reasoning and Planning for LLMs at ICLR 2025 @iclr_conf has received an overwhelming number of submissions! We are looking for reviewers to help ensure a high-quality selection process. 🔹 Max 2 papers per reviewer 🔹 Review deadline:…


We are excited to announce that our workshop will be held on April 28 in Singapore. Due to numerous requests for extensions, we have decided to extend the submission deadline by 4 days to February 6 (AoE). We look forward to receiving your submissions and can't wait to see you at…

ZhiyuanCS's tweet image. We are excited to announce that our workshop will be held on April 28 in Singapore. Due to numerous requests for extensions, we have decided to extend the submission deadline by 4 days to February 6 (AoE). We look forward to receiving your submissions and can't wait to see you at…

Zhiyuan hat repostet

🚀 Excited to announce our World Models: Understanding, Modelling and Scaling Workshop at #ICLR2025! 🎉 Keynote speakers, panellists, and submission guidelines are live now! Check out: 👉 sites.google.com/view/worldmode… Join us as we explore World Understanding, Sequential Modelling,…

Mengyue_Yang_'s tweet image. 🚀 Excited to announce our World Models: Understanding, Modelling and Scaling Workshop at #ICLR2025! 🎉

Keynote speakers, panellists, and submission guidelines are live now! Check out:
👉 sites.google.com/view/worldmode…

Join us as we explore World Understanding, Sequential Modelling,…

Our poster presentation at #NeurIPS2024 will take place today from 11:00 AM to 2:00 PM in West Ballroom A-D, Poster #7004. We warmly welcome you to stop by and engage with us!

How do LLMs conduct reasoning and planning given partial information with uncertainty? Whether they can proactively ask questions to improve decision-making? In joint work with UW, NTU, Yale and UCL, we introduce the UoT method, which boosts the information-seeking and…

ZhiyuanCS's tweet image. How do LLMs conduct reasoning and planning given partial information with uncertainty? Whether they can proactively ask questions to improve decision-making?

In joint work with UW, NTU, Yale and UCL, we introduce the UoT method, which boosts the information-seeking and…
ZhiyuanCS's tweet image. How do LLMs conduct reasoning and planning given partial information with uncertainty? Whether they can proactively ask questions to improve decision-making?

In joint work with UW, NTU, Yale and UCL, we introduce the UoT method, which boosts the information-seeking and…
ZhiyuanCS's tweet image. How do LLMs conduct reasoning and planning given partial information with uncertainty? Whether they can proactively ask questions to improve decision-making?

In joint work with UW, NTU, Yale and UCL, we introduce the UoT method, which boosts the information-seeking and…


Loading...

Something went wrong.


Something went wrong.