I am a PhD student studying computer science at the University of California, Santa Cruz, advised by Prof. Chenguang Wang and very fortunate to closely work with Prof. Dawn Song. I am also a Visiting Researcher at Scale AI. Previously, I completed my Bachelor’s degree in Data Science from the Mathematics Department at Washington University in St. Louis, where I was advised by Prof. Ulugbek Kamilov and graduated with Highest Distinction. You can find my full CV here.

My research focuses on developing methods to better interpret and ensure the safety and performance of large language models (LLMs) and LLM agents.

Research Interests: LLM Interpretability, Alignment & Safety, Agentic AI

📢 Announcements

  • May 2026: 3 works accepted to ICML 2026!
  • April 2026: Peer preservation and agent security formalization now out!
  • January 2026: RepIt is accepted to ICLR 2026!
  • December 2025: Our workshop on Agent Safety is accepted to ICLR - see you all in Brazil! I will be in NeurIPS this week as well.
  • September 2025: RepIt and SteeringSafety are now on arXiv!
  • August 2025: I will be moving with my advisor to UCSC to continue my PhD! Additionally, AgentVigil was accepted to EMNLP 2025!
  • March 2025: Excited to share COSMIC accepted to ACL 2025! 🎉
Formal Security Model framework

A Framework for Formalizing LLM Agent Security

Vincent Siu, Jingxuan He, Kyle Montgomery, Zhun Wang, Neil Gong, Chenguang Wang, Dawn Song

arXiv

Paper (arXiv) |

Peer Preservation

Peer Preservation in Frontier Models

Yujin Potter*, Nick Crispino*, Vincent Siu, Chenguang Wang, Dawn Song

ICML 2026

Paper | X Announcement (Tweet)

Find the full list of publications here

📰 Media Coverage

🤝 Service & Engagement

Open Source Software

  • MassGen — Contributor
    ⭐ 900+ GitHub stars