Compositional sparsity of learnable functions

Type: Article
Publication Date: 2024-05-15
Citations: 3
DOI: https://doi.org/10.1090/bull/1820

Abstract

Neural networks have demonstrated impressive success in various domains, raising the question of what fundamental principles underlie the effectiveness of the best AI systems and quite possibly of human intelligence. This perspective argues that compositional sparsity, or the property that a compositional function have “few” constituent functions, each depending on only a small subset of inputs, is a key principle underlying successful learning architectures. Surprisingly, all functions that are efficiently Turing computable have a compositional sparse representation. Furthermore, deep networks that are also sparse can exploit this general property to avoid the “curse of dimensionality”. This framework suggests interesting implications about the role that machine learning may play in mathematics.

Locations

  • DSpace@MIT (Massachusetts Institute of Technology)
  • Bulletin of the American Mathematical Society

Ask a Question About This Paper

Summary

Login to see paper summary

Similar Works

Action Title Date Authors
The mathematics of artificial intelligence 2023-12-15 Gitta Kutyniok
The Mathematics of Artificial Intelligence 2022-01-01 Gitta Kutyniok
+
Learnable Functions 1977-01-01 Andrzej Ehrenfeucht Jan Mycielski
Understanding Boolean Function Learnability on Deep Neural Networks: PAC Learning Meets Neurosymbolic Models 2020-01-01 Anderson Rocha Tavares Pedro Avelar João Marcos Flach Márcio Nicolau Luís C. Lamb Moshe Y. Vardi
The No Free Lunch Theorem, Kolmogorov Complexity, and the Role of Inductive Biases in Machine Learning 2023-01-01 Micah Goldblum Marc Finzi Keefer Rowan Andrew Gordon Wilson
Learning Nonlinearity of Boolean Functions: An Experimentation with Neural Networks 2025-02-03 S. Sai Kalyan Ranga Nandish Chattopadhyay Anupam Chattopadhyay
The Mathematics of Artificial Intelligence 2025-01-15 Gabriel Peyré
+
Formulas for Learning 1967-04-29
Fourier Circuits in Neural Networks: Unlocking the Potential of Large Language Models in Mathematical Reasoning and Modular Arithmetic 2024-02-12 Jiuxiang Gu Chenyang Li Yingyu Liang Zhenmei Shi Zhao Song Tianyi Zhou
Approximation smooth and sparse functions by deep neural networks without saturation 2020-01-01 Xia Liu
When does compositional structure yield compositional generalization? A kernel theory 2024-05-25 Samuel Lippl Kim Stachenfeld
Learning Reductions that Really Work 2015-01-01 Alina Beygelzimer Hal Daumé John Langford Paul Mineiro
Learning Neural Networks with Sparse Activations 2024-06-25 Pranjal Awasthi Nishanth Dikkala Pritish Kamath Raghu Meka
Feature emergence via margin maximization: case studies in algebraic tasks 2023-01-01 Depen Morwani Benjamin Edelman Costin-Andrei Oncescu Rosie Zhao Sham M. Kakade
The power of deeper networks for expressing natural functions 2017-01-01 David Rolnick Max Tegmark
Can Neural Networks Do Arithmetic? A Survey on the Elementary Numerical Skills of State-of-the-Art Deep Learning Models 2024-01-15 Alberto Testolin
Arithmetic Circuits, Structured Matrices and (not so) Deep Learning 2022-12-17 Atri Rudra
Recurrent Convolutional Neural Networks Learn Succinct Learning Algorithms 2022-01-01 Surbhi Goel Sham M. Kakade Adam Tauman Kalai Cyril Zhang
Seeing is Believing: Brain-Inspired Modular Training for Mechanistic Interpretability 2023-01-01 Ziming Liu Eric Gan Max Tegmark
+
Neural Power Units 2020-06-02 Niklas Heim Tomáš Pevný Václav Šmídl