Spotlight Poster
Feature learning from non-Gaussian inputs: the case of Independent Component Analysis in high dimensions
Fabiola Ricci · Lorenzo Bardone · Sebastian Goldt
West Exhibition Hall B2-B3 #W-812
How do neural networks learn to "see"? Unlike classical machine learning methods, neural networks automatically learn image-processing filters from data -- a key advantage over classical methods, but one that remains poorly understood. The main difficulty for the analysis is that neural networks exploit complex patterns that cannot be captured by simple averages or pair-wise relations between pixels, while existing theories can at most account for pair-wise relations, captured by Gaussian distributions.In this paper, we investigate feature learning by analysing a simpler method, Independent Component Analysis (ICA). ICA seeks to find the most non-Gaussian projections of inputs -- and finds similar filters as neural networks.Comparing the two most popular algorithms for ICA when inputs are large, we find that FastICA requires a lot more data than is theoretically required, while stochastic gradient descent (with a small modification) only requires the theoretical minimum of samples for learning. Since we make some simplifying assumptions on the data, we corroborate our findings in experiments with real images.Our results improve our understanding of how to learn efficiently from the complex correlations in real data, and we provide new technical tools for future analysis.