Breaking the bonds of generative artificial intelligence by minimizing the maximum entropy

Type: Preprint
Publication Date: 2025-02-18
Citations: 0
DOI: https://doi.org/10.48550/arxiv.2502.13287

Abstract

The emergence of generative artificial intelligence (GenAI), comprising large language models, text-to-image generators, and AI algorithms for medical drug and material design, had a transformative impact on society. However, despite an initial exponential growth surpassing Moore's law, progress is now plateauing, suggesting we are approaching the limits of current technology. Indeed, these models are notoriously data-hungry, prone to overfitting, and challenging to direct during the generative process, hampering their effective professional employment. To cope with these limitations, we propose a paradigm shift in GenAI by introducing an ab initio method based on the minimal maximum entropy principle. Our approach does not fit the data. Instead, it compresses information in the training set by finding a latent representation parameterized by arbitrary nonlinear functions, such as neural networks. The result is a general physics-driven model, which is data-efficient, resistant to overfitting, and flexible, permitting to control and influence the generative process. Benchmarking shows that our method outperforms variational autoencoders (VAEs) with similar neural architectures, particularly on undersampled datasets. We demonstrate the methods effectiveness in generating images, even with limited training data, and its unprecedented capability to customize the generation process a posteriori without the need of any fine-tuning or retraining.

Locations

  • arXiv (Cornell University)

Ask a Question About This Paper

Summary

This paper presents a novel approach to generative artificial intelligence (GenAI) based on the “minimal maximum entropy” principle. The significance lies in addressing limitations of current GenAI models, which are data-hungry, prone to overfitting, and difficult to control.

Key innovations:

  • Ab initio method: Instead of directly fitting the data, the method compresses information by finding a latent representation using arbitrary nonlinear functions (e.g., neural networks).
  • Physics-driven model: This approach results in a general model that is data-efficient, resistant to overfitting, and flexible for controlling the generative process.
  • Minimal maximum entropy principle: This principle is used to find the most unbiased probability distribution consistent with given constraints, ensuring no unwarranted assumptions.
  • Customization: The method allows for customizing the generation process a posteriori without fine-tuning or retraining.

Prior ingredients needed:

  • Generative Artificial Intelligence (GenAI): Models capable of algorithmically producing novel data resembling training data (e.g., large language models, text-to-image generators).
  • Maximum Entropy Principle: A guiding principle to assign probabilities to events by selecting the most unbiased probability distribution consistent with given constraints.
  • Variational Autoencoders (VAEs): A type of generative model that learns a latent representation of the data.
  • Neural Networks: A set of algorithms, modeled loosely after the human brain, that are designed to recognize patterns.
  • Principal Component Analysis (PCA): The algorithm used to reduce the dimensionality of a data set.
  • Training dataset: Used as the dataset of real data to create and then tune to.

Similar Works

Action Title Date Authors
Hitchhiker's guide on Energy-Based Models: a comprehensive review on the relation with other generative models, sampling and statistical physics 2024-06-19 Davide Carbone
On the Challenges and Opportunities in Generative AI 2024-02-28 Laura Manduchi Kushagra Pandey Robert Bamler Ryan Cotterell Sina Däubener Sophie Fellenz Asja Fischer Thomas Gärtner Matthias Kirchler Marius Kloft
Generative artificial intelligence for computational chemistry: a roadmap to predicting emergent phenomena 2024-09-04 Pratyush Tiwary Lukas Herron Richard S. John Suemin Lee Disha Sanwal Ruiyu Wang
Self-Supervised Generative Models for Crystal Structures 2023-01-01 Fangze Liu Zhaotao Chen Tianyi Liu Yu Lin Joshua J. Turner Chunjing Jia
A survey on Variational Autoencoders from a GreenAI perspective 2021-01-01 Andrea Asperti Daniele Evangelista Elena Loli Piccolomini
Rethinking Molecular Design: Integrating Latent Variable and Auto-Regressive Models for Goal Directed Generation 2024-08-19 Heath Arthur-Loui Amina Mollaysa Michael Krauthammer
Dual Space Training for GANs: A Pathway to Efficient and Creative Generative Models 2024-10-21 Beka Modrekiladze
Self-supervised generative models for crystal structures 2024-08-06 Fangze Liu Zhantao Chen Tianyi Liu Ruyi Song Yu Lin Joshua J. Turner Chunjing Jia
+
A Generic and Model-Agnostic Exemplar Synthetization Framework for Explainable AI 2020-06-06 Antonio Bărbălau Adrian Cosma Radu Tudor Ionescu Marius Popescu
Physics-Integrated Variational Autoencoders for Robust and Interpretable Generative Modeling 2021-01-01 Naoya Takeishi Alexandros Kalousis
Physics-Integrated Variational Autoencoders for Robust and Interpretable Generative Modeling 2021-02-25 Naoya Takeishi Alexandros Kalousis
Diffusion Models for Molecules: A Survey of Methods and Tasks 2025-02-13 Liang Wang Chao Song Zhiyuan Liu Rong Yu Qiang Liu Shu Wu
Study of Deep Generative Models for Inorganic Chemical Compositions 2019-01-01 Yoshihide Sawada Koji Morikawa Mikiya Fujii
A Generic and Model-Agnostic Exemplar Synthetization Framework for Explainable AI 2020-01-01 Antonio Bărbălau Adrian Cosma Radu Tudor Ionescu Marius Popescu
Kernel-Elastic Autoencoder for Molecular Design 2023-01-01 Haote Li Yu Shee Brandon Allen Federica Maschietto Víctor S. Batista
Variational Potential Flow: A Novel Probabilistic Framework for Energy-Based Generative Modelling 2024-07-21 Junn Yong Loo Michelle Adeline Arghya Pal Vishnu Monn Baskaran Chee-Ming Ting Raphaël C.‐W. Phan
Generative Artificial Intelligence: A Systematic Review and Applications 2024-05-17 Sandeep Singh Sengar Affan Bin Hasan Sanjay Kumar Fiona Carroll
When Geoscience Meets Generative AI and Large Language Models: Foundations, Trends, and Future Challenges 2024-01-25 Abdenour Hadid Tanujit Chakraborty D. Busby
Con-CDVAE: A method for the conditional generation of crystal structures 2024-03-19 Cai-Yuan Ye Hongming Weng Quan-Sheng Wu
Con-CDVAE: A method for the conditional generation of crystal structures 2024-04-29 Cai-Yuan Ye Hongming Weng Quansheng Wu

Cited by (0)

Action Title Date Authors

Citing (0)

Action Title Date Authors