LESSWRONG
LW

Interpretability (ML & AI)Machine Learning (ML)AI
Frontpage

8

Mechanism for feature learning in neural networks and backpropagation-free machine learning models

by Matt Goldenberg
19th Mar 2024
1 min read
1

8

This is a linkpost for https://www.science.org/doi/10.1126/science.adi5639
Interpretability (ML & AI)Machine Learning (ML)AI
Frontpage

8

Mechanism for feature learning in neural networks and backpropagation-free machine learning models
10mishka
New Comment
1 comment, sorted by
top scoring
Click to highlight new comments since: Today at 6:40 AM
[-]mishka1y100

I don't see a non-paywalled version of this paper, but here are two closely related preprints by the same authors:

"Mechanism of feature learning in deep fully connected networks and kernel machines that recursively learn features", https://arxiv.org/abs/2212.13881

"Mechanism of feature learning in convolutional neural networks", https://arxiv.org/abs/2309.00570

Reply
Moderation Log
More from Matt Goldenberg
View more
Curated and popular this week
1Comments

Understanding how neural networks learn features, or relevant patterns in data, for prediction is necessary for their reliable use in technological and scientific applications. In this work, we presented a unifying mathematical mechanism, known as Average Gradient Outer Product (AGOP), that characterized feature learning in neural networks. We provided empirical evidence that AGOP captured features learned by various neural network architectures, including transformer-based language models, convolutional networks, multi-layer perceptrons, and recurrent neural networks. Moreover, we demonstrated that AGOP, which is backpropagation-free, enabled feature learning in machine learning models, such as kernel machines, that apriori could not identify task-specific features. Overall, we established a fundamental mechanism that captured feature learning in neural networks and enabled feature learning in general machine learning models.