Ben's headshot

Hi, I'm Ben Hoover

I'm an AI Researcher studying memory

Understanding AI foundation models from the perspective of large Associative Memories.

I am a Machine Learning PhD student at Georgia Tech advised by Polo Chau and an AI Research Engineer with IBM Research. My research focuses on building more interpretable and parameter efficient AI by rethinking the way we train and build deep models, taking inspiration from Associative Memories and Hopfield Networks. I like to visualize what happens inside AI models.

News

Sep 2025
๐ŸŽ‰ Dense Associative Memory with Epanechnikov energy accepted as ๐Ÿ…Spotlight Poster (top 3% of submissions) to NeurIPS'25 main conference. See you in San Diego โœˆ๏ธ!
Sep 2025
๐Ÿ“ข Our Associative Memory Tutorial is accepted at AAAI 2026 in Singapore. See you there!
Aug 2025
๐ŸŽ‰ Dense Associative Memory with Epanechnikov energy accepted to the MemVis Workshop @ICCV'25. See you in Hawaii ๐ŸŒบ!
Aug 2025
๐Ÿ’ฌ I provide feedback for a Nature article on the risks of AI plagiarism.
July 2025
๐Ÿ“ข Diffusion Explainer is highlighted by Quanta Magazine's "How Physics Made Modern AI Possible"
July 2025
โœˆ๏ธ I will be at ICML in Vancouver presenting our Associative Memory Tutorial and ConceptAttention (oral ๐Ÿ†). Feel free to reach out!
See more...

Research Highlights

Thumbnail for Epanechnikov DenseAM

Epanechnikov DenseAM

Dense Associative Memories can produce emergent memories when we replace the Gaussian kernel with a KDE-optimal Epanechnikov kernel.
Thumbnail for Tutorial on Associative Memory

Tutorial on Associative Memory

Energy-based Associative Memory transformed the field of AI, but it is hardly understood. We present a birds eye view of Associative Memory, beginning with the invention of the Hopfield Network and concluding with modern, dense storage versions that have strong connections to Transformers and kernel methods.
Thumbnail for ConceptAttention

ConceptAttention

We discover how to extract highly salient features from the learned representations of Diffusion Transformers, using our technique to segment images by semantic concept. We name our technique ConceptAttention, which outperforms all prior methods on single- and multi-class classification.
Thumbnail for DenseAMs meet Random Features

DenseAMs meet Random Features

DenseAMs can store an exponential number of memories, but each memory adds new parameters to the energy function. We propose a novel, "Distributed representation for DenseAMs" (DrDAM) that allows us to add new memories without increasing the total number of parameters.
Thumbnail for Transformer Explainer

Transformer Explainer

Transformers are the most powerful AI innovation of the last decade. Learn how they work by interacting with every mechanic from the comfort of your web browser. Taught in Georgia Tech CSE6242 Data and Visual Analytics (typically 250-300 students per semester).
Thumbnail for Memory in Plain Sight

Memory in Plain Sight

We are the first work to discover that diffusion models perform memory retrieval in their denoising dynamics.
Thumbnail for Energy Transformer

Energy Transformer

We derive an Associative Memory inspired by the famous Transformer architecture, where the forward pass through the model is memory retrieval by energy descent.
Thumbnail for Diffusion Explainer

Diffusion Explainer

Diffusion models are complicated. We break down Stable Diffusion and explain each component of the model visually. Taught in Georgia Tech CSE6242 Data and Visual Analytics (typically 250-300 students per semester).
Thumbnail for HAMUX

HAMUX

We invent a software abstraction around "synapses" and "neurons" to assemble energy functions of complicated Associative Memories, where memory retrieval is performed through autograd.
Thumbnail for RXNMapper

RXNMapper

We discover that Transformers trained on chemical reactions learn, on their own, how atoms physically rearrange.