Yong Zheng-Xin

Current: CS PhD @ Brown University, Astra Fellow
Past: Research Scientist Intern @ Meta AI, Research Collaborator @ Cohere Labs

prof_pic.jpg

I am a final-year PhD student at Brown University advised by Prof. Stephen Bach. I am fortunate to have interned/collaborated at Meta (GenAI, FAIR) and Cohere Labs, and my research is funded by Open Philanthropy grant for technical AI safety.

I work on making AI systems safe and helpful for everyone around the world. My recent research focuses on post-training, especially on reasoning and safety alignment. In particular, I work on understanding surprising properties for the reasoning chain-of-thoughts (CoTs), such as

  • Cross-lingual reasoning through test-time scaling (preprint).
  • Self-jailbreaking where models try to reason themselves out of safety guardrail after benign reasoning training (preprint)
  • Being able to predict safety outcomes before models finish thinking (preprint).

A huge part of my previous research was on multilingual LLMs and speech models, especially on their alignment and capability on low-resource languages.

Actively seeking full-time research roles in industry.


Featured work/preprints (see all)

  1. Yik Siu Chan* ,  Zheng-Xin Yong* ,  and  Stephen H. Bach
    arxiv preprint, 2025
  2. Zheng-Xin Yong ,  M. Farid Adilazuarda ,  Jonibek Mansurov , and 7 more authors
    arxiv preprint, 2025
  3. Ahmet Üstün* ,  Viraat Aryabumi* ,  Zheng-Xin Yong* , and 14 more authors
    ACL, 2024 (Best Paper Award)
  4. Zheng-Xin Yong ,  Cristina Menghini ,  and  Stephen Bach
    NeurIPS Workshop: Socially Responsible Language Modelling Research (SoLaR) , 2023 (Best Paper Award)