Home
My Research
Talks
Publications
Teaching
Outreach
Contact
Light
Dark
Automatic
Structural Pruning
How a student becomes a teacher: learning and forgetting through Spectral methods
This study explores the teacher-student paradigm in machine learning, focusing on overparameterized student networks trained by fixed teacher networks. It introduces a new optimization scheme using spectral representation of linear information transfer between layers. This approach allows identifying a stable student substructure that mirrors the teacher’s complexity. The method shows that pruning unimportant nodes, based on optimized eigenvalues, does not degrade performance, indicating a second-order phase transition with universality traits in neural network training.
Lorenzo Giambagli
,
Lorenzo Buffoni
,
Lorenzo Chicchi
,
Duccio Fanelli
PDF
Cite
Source Document
Spectral pruning of fully connected layers
Training neural networks in spectral space focuses on optimizing eigenvalues and eigenvectors instead of individual weights, allowing effective implicit bias that node enables pruning without sacrificing performance.
Lorenzo Buffoni
,
Enrico Civitelli
,
Lorenzo Giambagli
,
Lorenzo Chicchi
,
Duccio Fanelli
PDF
Cite
Source Document
Training of sparse and dense deep neural networks: Fewer parameters, same performance
This study presents a variant of spectral learning for deep neural networks, where adjusting two sets of eigenvalues for each layer mapping significantly enhances network performance with fewer trainable parameters. This method, inspired by homeostatic plasticity, offers a computationally efficient alternative to conventional training, achieving comparable results with a simpler parameter setup. It also enables the creation of sparser networks with impressive classification abilities.
Lorenzo Chicchi
,
Lorenzo Giambagli
,
Lorenzo Buffoni
,
Timoteo Carletti
,
Marco Ciavarella
,
Duccio Fanelli
PDF
Cite
Source Document
Cite
×