Zilong Zheng

Email: z.zheng[at]ucla[dot]edu

I received my Ph.D. degree (21’) from the Department of Computer Science at University of California, Los Angeles (UCLA). My research interests lie in the intersection of statistical machine learning, natural language processing and cognition. Current research themes include:

  • Human AI Alignment: Building interactive models that align with human values and social norms.
  • Long-context Language Models: Efficient training and inference of long-context language models.
  • Generative Modeling: Statistical generative modeling (e.g. EBMs, diffusions) on high-dimensional data.

I am always looking for self-motivated students and long-term collaborators. Please contact me if you have excellent background or share similar research interests with me.

News

2024/12 I will co-host 1st workshop on Large Language Models and Structure Modeling. Stay tuned :fire:.
2024/12 Diver-CT is accepted to AAAI’25. Congratulations to Andrew!
2024/09 Two papers on long context window extension and situated inductive reasoning are accepted to NeurIPS’24. Congratulations to Tong and Xiaojuan!
2024/09 Two papers on video understanding and sentence representation are accepted to EMNLP’24 Main. Congratulations to Yuxuan and Ziyong!
2024/07 One paper on ToM for dialogue modeling is accepted for an Oral presentation at SIGDial’24. Congratulations to Shuwen!

Selected Publications

*: Equal contribution, : Corresponding author
  1. DiveR-CT: Diversity-enhanced Red Teaming Large Language Model Assistants with Relaxing Constraints AAAI'25
    Andrew Zhao, Quentin Xu, Matthieu Liu, Shenzhi Wang, Yong-jin Liu, Zilong Zheng, and Gao Huang, in AAAI, 2025.
  2. Mars: Situated Inductive Reasoning in an Open-World Environment NeurIPS'24
    Xiaojuan Tang, Jiaqi Li, Yitao Liang, Muhan Zhang, and Zilong Zheng, in NeurIPS D&B Track, 2024.
  3. An Efficient Recipe for Long Context Extension via Middle-Focused Positional Encoding NeurIPS'24
    Tong Wu, Yanpeng Zhao, and Zilong Zheng, in NeurIPS, 2024.
  4. How to Synthesize Text Data without Model Collapse?
    Xuekai Zhu, Daixuan Cheng, Hengli Li, Kaiyan Zhang, Ermo Hua, Xingtai Lv, Ning Ding, Zhouhan Lin, Zilong Zheng, and Bowen Zhou.
  5. Sparser is Faster and Less is More: Efficient Sparse Attention for Long-Range Transformers
    Chao Lou, Zixia Jia, Zilong Zheng, and Kewei Tu.
  6. In-Context Editing: Learning Knowledge from Self-Induced Distributions
    Siyuan Qi, Bangcheng Yang, Kailin Jiang, Xiaobo Wang, Jiaqi Li, Yifan Zhong, Yaodong Yang, and Zilong Zheng.
  7. In situ bidirectional human-robot value alignment ScienceRobotics
    Luyao Yuan*, Xiaofeng Gao*, Zilong Zheng*, Mark Edmonds, Ying Nian Wu, Federico Rossano, Hongjing Lu, Yixin Zhu, and Song-Chun Zhu, Science Robotics, 2022.
  8. Patchwise Generative ConvNet: Training Energy-Based Models from a Single Natural Image for Internal Learning Oral CVPR'21
    Zilong Zheng, Jianwen Xie, and Ping Li, in CVPR, 2021.
  9. Reasoning Visual Dialogs with Structural and Partial Observations Oral CVPR'19
    Zilong Zheng*, Wenguan Wang*, Siyuan Qi*, and Song-Chun Zhu, in CVPR, 2019.
  10. Learning Descriptor Networks for 3D Shape Synthesis and Analysis Oral CVPR'18
    Jianwen Xie*, Zilong Zheng*, Ruiqi Gao, Wenguan Wang, Song-Chun Zhu, and Ying Nian Wu, in CVPR, 2018.