Skip to main content

Post #6: Traditional Statistics: Descriptive and Inferential in Big Data

Two types of Traditional statistics in Big data include Descriptive and Inferential. Descriptive means averages and working on sets of numbers. Descriptive statistics is a type of statistic in which a data set is summarised and the characteristics are described. This descriptive data is usually displayed through the use of tables, charts etc. but is most commonly reported as a measure of a central tendency. A central tendency is a typical value for a distribution, it is also been known to be called a location or centre of the distribution. The arithmetic mean is the most common measure of a central tendency, this is the median and the mode. The mean is the average of all the values, the median is the exact middle of the data set while the mode is the most frequent value in the data set. 

The goal of traditional statistics is analysing and summarising data, providing tight assumptions about the problem and data distributions as well as using conservative techniques and approaches.  

Comments

Popular posts from this blog

FutureLearn Week 3: Post 1 of 3

Citizen science can be applied to any and all kinds of scientific observations which require scientific monitoring data. Weather phenomena, climate change impacts in the environment, monitoring of local air quality and also monitoring the effects and impact of climate change on different species of animal. Citizen science for climate simulation and ecology can be incredibly helpful in monitoring many things such as monitoring microplastics in water so that people can be more aware of what products contribute to this and lead efforts to reduce or change the process in which manufacturers produce these products to eliminate it from water sources. 

FutureLearn Week 2: Post 3 of 4

Two of the biggest challenges of big data is Analysing and Visualising the data. Firstly with analysing the data, the size of big data files can sometimes be substantial, there are many things that must be considered before downloading the data, for example the file size, how long the data file will take to download, will all of it be necessary or will part of the file suffice and is there enough storage space within the system itself. Visualisation is way to represent the data in a way that is easier to understand such as word clouds and things of the like. This will aid users in seeing the prominent and key terms from the analysis of the data sets. The first step after downloading the data would be to quality check it to ensure that each field had the appropriate data types in each field and to ensure that the user understood the meaning of each field. Keeping a copy of the original data would be essential as well as each documented version change for each stage of visualisation....