Decomposing the Dark Matter of Sparse Autoencoders.
Joshua Engels, Logan Smith, and Max Tegmark.
Preprint |
Code |
Twitter
Efficient Dictionary Learning with Switch Sparse Autoencoders.
Anish Mudide, Joshua Engels, Eric J Michaud, Max Tegmark, and Christian Schroeder de Witt.
Preprint |
Code |
Twitter
Not All Language Model Features Are Linear.
Joshua Engels, Eric J. Michaud, Isaac Liao, Wes Gurnee, and Max Tegmark.
Preprint |
Code |
Twitter |
Talk
Approximate Nearest Neighbor Search with Window Filters.
Joshua Engels, Benjamin Landrum, Shangdi Yu, Laxman Dhulipala, and Julian Shun.
ICML 2024.
Paper |
Code
DESSERT: An Efficient Algorithm for Vector Set Search with Vector Set Queries.
Joshua Engels, Benjamin Coleman, Vihan Lakshman, and Anshumali Shrivastava
NeurIPS 2023.
Paper |
Code |
Blog Post
Practical Near Neighbor Search via Group Testing.
Joshua Engels*, Benjamin Coleman*, and Anshumali Shrivastava
NeurIPS 2021: Spotlight talk - top 3%
Paper |
Talk |
Code
SAE Probing: What is it good for? Absolutely something! (2024) - We examine whether SAE probes are more data efficient and robust than activation probes.
Examining the Interaction of Interpretable Features and Training Dynamics in Othello-GPT (2023) - We experiment with promoting linear features while training Othello-GPT.