BIG DATA AND THE CENTRAL LIMIT THEOREM: A STATISTICAL LEGEND

Authors

  • S. Allende-Alonso Universidad de La Habana
  • C. N. Bouza-Herrera Universidad de La Habana
  • S. E. H. Rizvi Agricultural Statistics, SKUAST- India
  • J. M. Sautto-Vallejo Universidad Autónoma de Guerrero

Keywords:

Big-Data, normality tests, asymptotic normality of means

Abstract

Nowadays we deal with Big-Data commonly. The users of statistics rely on having a large sample size n for using the statistical methods based on normality. Usual inference methods are typically based on considering the Normal as the limit distributions of the sample mean for a large n. With large enough sample sizes (> 30 or 40), the violation of the normality assumption should not cause major problems. This fact implies that we can use parametric procedures even when the data are not normally distributed. Al least a goodness-of-fit test must be performed for accepting whether normality is valid or not. Monte Carlo (MC) techniques are used for selecting independent random samples of populations of means of three variables of importance in web network management. Different tests are performed to establish the acceptance of the normality. We did not find reliable results even for samples of size 10 000

Downloads

Download data is not yet available.

Downloads

Published

2023-04-11

How to Cite

Allende-Alonso, S., Bouza-Herrera, C. N., Rizvi, S. E. H., & Sautto-Vallejo, J. M. (2023). BIG DATA AND THE CENTRAL LIMIT THEOREM: A STATISTICAL LEGEND. Investigación Operacional, 40(1). Retrieved from https://revistas.uh.cu/invoperacional/article/view/2694

Similar Articles

1 2 3 4 5 6 7 8 9 10 > >> 

You may also start an advanced similarity search for this article.