Luxi (Lucy) He

prof_pic.jpg

Hi and welcome to Lucy’s homepage! I’m currently a third-year CS Ph.D. student at Princeton University, where I’m fortunate to be co-advised by Prof. Danqi Chen and Prof. Peter Henderson. My current research focuses on understanding language models and improving their alignment and safety. Recently, I’ve been particularly interested in building better human-LM alignment and human-agent collaboration from first principles. I’m also interested in the impact of data across the language model lifecycle — from understanding and predicting how data shapes training to improving model reliability. Motivated by real-world impact and a desire to bridge the gap between technology and policy, I aim to bring insights from both domains into my research.

Before Princeton, I obtained my Bachelor’s degree from Harvard with Highest Honors in Computer Science & Mathematics and a concurrent Master’s in Applied Math.

Outside of research, I’m a singer, dancer, photographer, and amateur food blogger.

Email: luxihe at princeton.edu

news

2025-12 Our workshop on Navigating and Addressing Data Problems for Foundation Models has been accepted to ICLR 2026! The workshop will take place on April 26th, 2026.
2025-10 Gave an oral presentation of our AudioLM evaluation paper at AIES 2025.
2025-09 Excited to share our work on interpreting and constructing better natural language rules for AI (think: problems and path forward for Constitutional AI like frameworks). Don’t miss the accompanying X thread, blog post, and policy brief!
2025-06 Started my internship at Google Research in Mountain View, CA.
2024-12 Attended NeurIPS in Vancouver! Presented CharXiv and gave an oral presentation at the EvalEval Workshop.

selected publications

  1. CAI_cover.png
    Statutory Construction and Interpretation for Artificial Intelligence
    Luxi He*, Nimra Nadeem*, Michel Liao, Howard Chen, Danqi Chen , and 2 more authors
    NeurIPS 2025 RegML Workshop (Oral); PNAS, 2026
  2. dataprophet_cover.png
    DataProphet: Demystifying Supervision Data Generalization in Multimodal LLMs
    Xuan Qi, Luxi He, Dan Roth, and Xingyu Fu
    ICLR, 2026
  3. audiolm_illustration.png
    The Model Hears You: Audio Language Model Deployments Should Consider the Principle of Least Privilege
    Luxi He*, Xiangyu Qi*, Michel Liao, Inyoung Cheong, Prateek Mittal , and 2 more authors
    AIES (Oral), 2025
  4. copycat_cover.png
    Fantastic Copyrighted Beasts and How (Not) to Generate Them
    Luxi He*, Yangsibo Huang*, Weijia Shi*, Tinghao Xie, Haotian Liu , and 5 more authors
    ICLR 2025; ICML GenLaw Workshop (Spotlight), 2025
  5. MeCo_cover.png
    Metadata Conditioning Accelerates Language Model Pre-training
    Tianyu Gao, Alexander Wettig, Luxi He, Yihe Dong, Sadhika Malladi , and 1 more author
    ICML, 2025
  6. benign_data_safety.png
    What is in Your Safe Data? Identifying Benign Data that Breaks Safety
    Luxi He*, Mengzhou Xia*, and Peter Henderson
    Conference on Language Modeling (COLM), ICLR Data Problems in Foundation Model (Best Paper), 2024
  7. charxiv_cover.png
    CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs
    Zirui Wang, Mengzhou Xia, Luxi He, Howard Chen, Yitao Liu , and 8 more authors
    NeurIPS Datasets & Benchmarks, 2024
  8. fairfront_cover.png
    Aleatoric and Epistemic Discrimination: Fundamental Limits of Fairness Interventions
    Hao Wang, Luxi He, Rui Gao, and Flavio Calmon
    In NeurIPS (Spotlight) , 2023