RANDOM VARIABLE GENERATION ALGORITHMS
Authors
Abstract
Generative algorithms are a rapidly developing field in artificial intelligence. The generation and retrieval of text, images, and video are becoming increasingly popular tools to assist in content creation. In recent years, training machine learning algorithms using big data has become one of the key tasks of modern science and practice. However, in a number of areas, the availability of truly large data sets is limited or impossible due to confidentiality, the high cost of collecting information, or the rarity of events. In such cases, synthetic data generation becomes particularly important, providing opportunities for data augmentation, modeling, and experimental evaluation without the risk of exposing sensitive information. Modern approaches to synthetic data rely both on statistical methods for generating random variables and on advanced generative models, such as Variational Autoencoders (VAE), Generative Adversarial Networks (GAN), and diffusion models, which demonstrate high accuracy and flexibility across various applied tasks. This paper provides a review of algorithms that generate discrete and continuous random variables. Special attention is given to models for sequential data, including TimeGAN, and RTSGAN, which enable effective modeling of time series in healthcare, financial forecasting, and autonomous systems. The article discusses the architectures of these models, their training methodologies, key innovations, and practical applications. Additionally, methods for evaluating the quality of synthetic data, their reliability, and ethical considerations are addressed. Based on an analysis of recent research, promising areas for the development of generative models aimed at improving interpretability, reliability, and specific focus on subject areas are identified.
Keywords
synthetic data, generative algorithms, generalised entropy method, neural networks, competitive generative models, variational autoencoders.
References
1. Azimi A., et.al. (2024), zGAN: An Outlier-focused Generative Adversarial Network For Realistic Synthetic Data Generation, Machine Learning, doi.org/10.48550/arXiv.2410.20808.
2. Diederik P. Kingma and Max Welling (2019), “An Introduction to Variational Autoencoders”, Foundations and Trends R in Machine Learning.
3. Hennig, P. (2012). Entropy Search for Information-Efficient Global Optimization. Journal of Machine Learning Research, 13, 1809–1837. https://jmlr.csail.mit.edu/papers/volume13/hennig12a /hennig12a.pdf.
4. Ian J.Goodfellow, et.al. (2014), Generative Adversarial Nets, Machine Learning, https://doi.org/10.48550/arXiv.1406.2661.
5. Jaynes, E.T. Information theory and statistical mechanics / E.T. Jaynes // Physical Reviews – 1957– N106 – P.620-630.
6. Kapur, J.N., Kesavan, H.K. Entropy optimization principles with applications, New York: Academic Press. – 1992, 360 p.
7. Laurens de Haan and Ana Ferreira. Extreme Value Theory. Springer New York, 2006.
8. Sasmita Manjari Nayak and Minakhi Rout. Impact of different outlier handling techniques on gan based hybrid bankruptcy prediction models. Indian Journal Of Science And Technology, 17:373–385, 1 2024.
9. Shamilov, A. (2007). Generalized entropy optimization problems and the existence of their solutions, Physica A: Statistical mechanics and its applications, 382, 465–472.
10. Shamilov, A. Generalized entropy optimization problems with finite moment function sets,/ A. Shamilov // Journal of Statistics and Management Systems –2010– N 13– P. 595-603.
11. Shannon, C.E. A mathematical theory of communications / C.E. Shannon // Bell System technical Journal – 1948 – N 27 – P.379–623.
12. Song, L. (2025). Adaptive generative adversarial maximum entropy inverse reinforcement learning. Information Sciences, Volume 695, pp.202-208.
Publish date
2026-03-30