Survival of the Fittest Representation: A Case Study with Modular Addition
Xiaoman Delores Ding*, Zifan Carl Guo*, Eric J. Michaud, Ziming Liu, Max Tegmark. In Mechanistic Interpretability Workshop at the ICML 2024. [arXiv][Twitter][Code]

Universal Neurons in GPT2 Language Models
Wes Gurnee, Theo Horsley, Zifan Carl Guo, Tara Rezaei Kheirkhah, Qinyi Sun, Will Hathaway, Neel Nanda, Dimitris Bertsimas. In TMLR. [arXiv][Twitter][Code]

Algorithmic progress in language models
Anson Ho, Tamay Besiroglu, Ege Erdil, David Owen, Robi Rahman, Zifan Carl Guo, David Atkinson, Neil Thompson, Jaime Sevilla. In Neurips 2024. [arXiv]

Measuring the Success of Diffusion Models at Imitating Human Artists
Stephen Casper*, Zifan Guo*, Shreya Mogulothu, Zachary Marinov, Chinmay Deshpande, Rui-Jie Yew, Zheng Dai, and Dylan Hadfield-Menell. Spotlighted in 2023 ICML Workshop on Generative AI and Law (GenLaw). (arXiv) (Twitter)

Enabling Transformers to Understand Low-Level Programs
Zifan Carl Guo, and William Moses. In 2022 IEEE High Performance Extreme Computing Conference (HPEC).


Modifications © Carl Guo 2023 & Tianyu Zhang 2021. Original source © R. Miles McCain 2020. Content is licensed CC BY-SA 4.0, a Free Culture License. The modified and original source codes are available under GPLv3.