I'm an undergraduate at UC Berkeley doing what I can to make AI go well!
In the past, I've worked on multi-turn jailbreaks, singular learning theory, and interpretability.
In my free time, I help organize the UC Berkeley AI Safety club, as well as some AI Safety related conferences.
Here's my resume for more info on what I'm up to!
I use this form for anonymous feedback/messages about how I can be better, and really appreciate people taking the time to fill it out.
I’m always up to chat. If you’re seeing this you should totally email me or reach out on twitter!
I also sometimes post here
Research Publications
Also see my Google Scholar.
- Emerging Vulnerabilities in Frontier Models: Multi-Turn Jailbreak Attacks
First co-author, advised by Kellin Pelrine (FAR ai).- 📍 NeurIPS SafeGenAI, NeurIPS Red Teaming GenAI
- Approximating the Local Learning Coefficient in Neural Networks: A Comparative Analysis of Power Series Expansion Orders
Advised by Lucius Bushnaq (Apollo).
Posts
Probably Not A Ghost Story
Making Little Simz Gorilla Interactive Music Video
Computer Apps I Recommend
subscribe via RSS