Basic Statistical Tools in Research and Data Analysis
Basic Statistical Tools in Research and Data Analysis
Karnataka, India
Address for correspondence: Dr. Zulfiqar Ali, Department of Anaesthesiology and Critical Care, Sheri
Kashmir Institute of Medical Sciences, Soura, Jammu and Kashmir, Srinagar, India. E-mail:
zulfiqaraliiii@yahoo.com
This is an open access article distributed under the terms of the Creative Commons Attribution-
NonCommercial-ShareAlike 3.0 License, which allows others to remix, tweak, and build upon the work non-
commercially, as long as the author is credited and the new creations are licensed under the identical terms.
Abstract
Statistical methods involved in carrying out a study include planning, designing, collecting data,
analysing, drawing meaningful interpretation and reporting of the research findings. The statistical
analysis gives meaning to the meaningless numbers, thereby breathing life into a lifeless data. The
results and inferences are precise only if proper statistical tests are used. This article will try to acquaint
the reader with the basic research tools that are utilised while conducting various studies. The article
covers a brief outline of the variables, an understanding of quantitative and qualitative variables and the
measures of central tendency. An idea of the sample size estimation, power analysis and the statistical
errors is given. Finally, there is a summary of parametric and non-parametric tests used for data
analysis.
Key words: Basic statistical tools, degree of dispersion, measures of central tendency, parametric tests
and non-parametric tests, variables, variance
INTRODUCTION
Statistics is a branch of science that deals with the collection, organisation, analysis of data and
drawing of inferences from the samples to the whole population.[1] This requires a proper design of the
study, an appropriate selection of the study sample and choice of a suitable statistical test. An adequate
knowledge of statistics is necessary for proper designing of an epidemiological study or a clinical trial.
Improper statistical methods may result in erroneous conclusions which may lead to unethical practice.
[2]
VARIABLES
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 1/14
2/23/2019 Basic statistical tools in research and data analysis
Variable is a characteristic that varies from one individual member of population to another individual.
[3] Variables such as height and weight are measured by some type of scale, convey quantitative
information and are called as quantitative variables. Sex and eye colour give qualitative information
and are called as qualitative variables[3] [Figure 1].
Figure 1
Classification of variables
Quantitative variables
Quantitative or numerical data are subdivided into discrete and continuous measurements. Discrete
numerical data are recorded as a whole number such as 0, 1, 2, 3,… (integer), whereas continuous data
can assume any value. Observations that can be counted constitute the discrete data and observations
that can be measured constitute the continuous data. Examples of discrete data are number of episodes
of respiratory arrests or the number of re-intubations in an intensive care unit. Similarly, examples of
continuous data are the serial serum glucose levels, partial pressure of oxygen in arterial blood and the
oesophageal temperature.
A hierarchical scale of increasing precision can be used for observing and recording the data which is
based on categorical, ordinal, interval and ratio scales [Figure 1].
Categorical or nominal variables are unordered. The data are merely classified into categories and
cannot be arranged in any particular order. If only two categories exist (as in gender male and female),
it is called as a dichotomous (or binary) data. The various causes of re-intubation in an intensive care
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 2/14
2/23/2019 Basic statistical tools in research and data analysis
unit due to upper airway obstruction, impaired clearance of secretions, hypoxemia, hypercapnia,
pulmonary oedema and neurological impairment are examples of categorical variables.
Ordinal variables have a clear ordering between the variables. However, the ordered data may not have
equal intervals. Examples are the American Society of Anesthesiologists status or Richmond agitation-
sedation scale.
Interval variables are similar to an ordinal variable, except that the intervals between the values of the
interval variable are equally spaced. A good example of an interval scale is the Fahrenheit degree scale
used to measure temperature. With the Fahrenheit scale, the difference between 70° and 75° is equal to
the difference between 80° and 85°: The units of measurement are equal throughout the full range of
the scale.
Ratio scales are similar to interval scales, in that equal differences between scale values have equal
quantitative meaning. However, ratio scales also have a true zero point, which gives them an additional
property. For example, the system of centimetres is an example of a ratio scale. There is a true zero
point and the value of 0 cm means a complete absence of length. The thyromental distance of 6 cm in
an adult may be twice that of a child in whom it may be 3 cm.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 3/14
2/23/2019 Basic statistical tools in research and data analysis
Table 1
Example of descriptive and inferential statistics
Descriptive statistics
The extent to which the observations cluster around a central location is described by the central
tendency and the spread towards the extremes is described by the degree of dispersion.
Measures of central tendency The measures of central tendency are mean, median and mode.[6] Mean
(or the arithmetic average) is the sum of all the scores divided by the number of scores. Mean may be
influenced profoundly by the extreme variables. For example, the average stay of organophosphorus
poisoning patients in ICU may be influenced by a single patient who stays in ICU for around 5 months
because of septicaemia. The extreme values are called outliers. The formula for the mean is
Mean,
where x = each observation and n = number of observations. Median[6] is defined as the middle of a
distribution in a ranked data (with half of the variables in the sample above and half below the median
value) while mode is the most frequently occurring variable in a distribution. Range defines the spread,
or variability, of a sample.[7] It is described by the minimum and maximum values of the variables. If
we rank the data and after ranking, group the observations into percentiles, we can get better
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 4/14
2/23/2019 Basic statistical tools in research and data analysis
information of the pattern of spread of the variables. In percentiles, we rank the observations into 100
equal parts. We can then describe 25%, 50%, 75% or any other percentile amount. The median is the
50th percentile. The interquartile range will be the observations in the middle 50% of the observations
about the median (25th -75th percentile). Variance[7] is a measure of how spread out is the distribution.
It gives an indication of how close an individual observation clusters about the mean value. The
variance of a population is defined by the following formula:
where σ2 is the population variance, X is the population mean, Xi is the ith element from the population
and N is the number of elements in the population. The variance of a sample is defined by slightly
different formula:
where s2 is the sample variance, x is the sample mean, xi is the ith element from the sample and n is the
number of elements in the sample. The formula for the variance of a population has the value ‘n’ as the
denominator. The expression ‘n−1’ is known as the degrees of freedom and is one less than the number
of parameters. Each observation is free to vary, except the last one which must be a defined value. The
variance is measured in squared units. To make the interpretation of the data simple and to retain the
basic unit of observation, the square root of variance is used. The square root of the variance is the
standard deviation (SD).[8] The SD of a population is defined by the following formula:
where σ is the population SD, X is the population mean, Xi is the ith element from the population and N
is the number of elements in the population. The SD of a sample is defined by slightly different
formula:
where s is the sample SD, x is the sample mean, xi is the ith element from the sample and n is the
number of elements in the sample. An example for calculation of variation and SD is illustrated in
Table 2.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 5/14
2/23/2019 Basic statistical tools in research and data analysis
Table 2
Example of mean, variance, standard deviation
Normal distribution or Gaussian distribution Most of the biological variables usually cluster around a
central value, with symmetrical positive and negative deviations about this point.[1] The standard
normal distribution curve is a symmetrical bell-shaped. In a normal distribution curve, about 68% of
the scores are within 1 SD of the mean. Around 95% of the scores are within 2 SDs of the mean and
99% within 3 SDs of the mean [Figure 2].
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 6/14
2/23/2019 Basic statistical tools in research and data analysis
Figure 2
Skewed distribution It is a distribution with an asymmetry of the variables about its mean. In a
negatively skewed distribution [Figure 3], the mass of the distribution is concentrated on the right of
Figure 1. In a positively skewed distribution [Figure 3], the mass of the distribution is concentrated on
the left of the figure leading to a longer right tail.
Figure 3
Inferential statistics
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 7/14
2/23/2019 Basic statistical tools in research and data analysis
In inferential statistics, data are analysed from a sample to make inferences in the larger collection of
the population. The purpose is to answer or test the hypotheses. A hypothesis (plural hypotheses) is a
proposed explanation for a phenomenon. Hypothesis tests are thus procedures for making rational
decisions about the reality of observed effects.
Probability is the measure of the likelihood that an event will occur. Probability is quantified as a
number between 0 and 1 (where 0 indicates impossibility and 1 indicates certainty).
In inferential statistics, the term ‘null hypothesis’ (H0 ‘H-naught,’ ‘H-null’) denotes that there is no
relationship (difference) between the population variables in question.[9]
Alternative hypothesis (H1 and Ha) denotes that a statement between the variables is expected to be
true.[9]
The P value (or the calculated probability) is the probability of the event occurring by chance if the null
hypothesis is true. The P value is a numerical between 0 and 1 and is interpreted by researchers in
deciding whether to reject or retain the null hypothesis [Table 3].
Table 3
P values with interpretation
If P value is less than the arbitrarily chosen value (known as α or the significance level), the null
hypothesis (H0) is rejected [Table 4]. However, if null hypotheses (H0) is incorrectly rejected, this is
known as a Type I error.[11] Further details regarding alpha error, beta error and sample size
calculation and factors influencing them are dealt with in another section of this issue by Das S et al.
[12]
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 8/14
2/23/2019 Basic statistical tools in research and data analysis
Table 4
Illustration for null hypothesis
PARAMETRIC AND NON-PARAMETRIC TESTS Numerical data (quantitative variables) that are
normally distributed are analysed with parametric tests.[13]
The assumption of normality which specifies that the means of the sample group are normally
distributed
The assumption of equal variance which specifies that the variances of the samples and of their
corresponding population are equal.
However, if the distribution of the sample is skewed towards one side or the distribution is unknown
due to the small sample size, non-parametric[14] statistical techniques are used. Non-parametric tests
are used to analyse ordinal and categorical data.
Parametric tests The parametric tests assume that the data are on a quantitative (numerical) scale, with
a normal distribution of the underlying population. The samples have the same variance (homogeneity
of variances). The samples are randomly drawn from the population, and the observations within a
group are independent of each other. The commonly used parametric tests are the Student's t-test,
analysis of variance (ANOVA) and repeated measures ANOVA.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 9/14
2/23/2019 Basic statistical tools in research and data analysis
Student's t-test
Student's t-test is used to test the null hypothesis that there is no difference between the means of the
two groups. It is used in three circumstances:
1. To test if a sample mean (as an estimate of a population mean) differs significantly from a given
population mean (this is a one-sample t-test)
where X1 − X2 is the difference between the means of the two groups and SE denotes the
standard error of the difference.
3. To test if the population means estimated by two dependent samples differ significantly (the
paired t-test). A usual setting for paired t-test is when measurements are made on the same
subjects before and after a treatment.
where d is the mean difference and SE denotes the standard error of this difference.
The group variances can be compared using the F-test. The F-test is the ratio of variances (var l/var 2).
If F differs significantly from 1.0, then it is concluded that the group variances differ significantly.
Analysis of variance
The Student's t-test cannot be used for comparison of three or more groups. The purpose of ANOVA is
to test if there is any significant difference between the means of two or more groups.
In ANOVA, we study two variances – (a) between-group variability and (b) within-group variability.
The within-group variability (error variance) is the variation that cannot be accounted for in the study
design. It is based on random differences present in our samples.
However, the between-group (or effect variance) is the result of our treatment. These two estimates of
variances are compared using the F-test.
where MSb is the mean squares between the groups and MSw is the mean squares within groups.
As with ANOVA, repeated measures ANOVA analyses the equality of means of three or more groups.
However, a repeated measure ANOVA is used when all variables of a sample are measured under
different conditions or at different points in time.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 10/14
2/23/2019 Basic statistical tools in research and data analysis
As the variables are measured from a sample at different points of time, the measurement of the
dependent variable is repeated. Using a standard ANOVA in this case is not appropriate because it fails
to model the correlation between the repeated measures: The data violate the ANOVA assumption of
independence. Hence, in the measurement of repeated dependent variables, repeated measures ANOVA
should be used.
Non-parametric tests
When the assumptions of normality are not met, and the sample means are not normally, distributed
parametric tests can lead to erroneous results. Non-parametric tests (distribution-free test) are used in
such situation as they do not require the normality assumption.[15] Non-parametric tests may fail to
detect a significant difference when compared with a parametric test. That is, they usually have less
power.
As is done for the parametric tests, the test statistic is compared with known values for the sampling
distribution of that statistic and the null hypothesis is accepted or rejected. The types of non-parametric
analysis techniques and the corresponding parametric analysis techniques are delineated in Table 5.
Table 5
Analogue of parametric and non-parametric tests
Median test for one sample: The sign test and Wilcoxon's signed rank test
The sign test and Wilcoxon's signed rank test are used for median tests of one sample. These tests
examine whether one instance of sample data is greater or smaller than the median reference value.
Sign test
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 11/14
2/23/2019 Basic statistical tools in research and data analysis
This test examines the hypothesis about the median θ0 of a population. It tests the null hypothesis H0 =
θ0. When the observed value (Xi) is greater than the reference value (θ0), it is marked as+. If the
observed value is smaller than the reference value, it is marked as − sign. If the observed value is equal
to the reference value (θ0), it is eliminated from the sample.
If the null hypothesis is true, there will be an equal number of + signs and − signs.
The sign test ignores the actual values of the data and only uses + or − signs. Therefore, it is useful
when it is difficult to measure the values.
There is a major limitation of sign test as we lose the quantitative information of the given data and
merely use the + or – signs. Wilcoxon's signed rank test not only examines the observed values in
comparison with θ0 but also takes into consideration the relative sizes, adding more statistical power to
the test. As in the sign test, if there is an observed value that is equal to the reference value θ0, this
observed value is eliminated from the sample.
Wilcoxon's rank sum test ranks all data points in order, calculates the rank sum of each sample and
compares the difference in the rank sums.
Mann-Whitney test
It is used to test the null hypothesis that two samples have the same median or, alternatively, whether
observations in one sample tend to be larger than observations in the other.
Mann–Whitney test compares all data (xi) belonging to the X group and all data (yi) belonging to the Y
group and calculates the probability of xi being greater than yi: P (xi > yi). The null hypothesis states
that P (xi > yi) = P (xi < yi) =1/2 while the alternative hypothesis states that P (xi > yi) ≠1/2.
Kolmogorov-Smirnov test
The two-sample Kolmogorov-Smirnov (KS) test was designed as a generic method to test whether two
random samples are drawn from the same distribution. The null hypothesis of the KS test is that both
distributions are identical. The statistic of the KS test is a distance between the two empirical
distributions, computed as the maximum absolute difference between their cumulative curves.
Kruskal-Wallis test
The Kruskal–Wallis test is a non-parametric test to analyse the variance.[14] It analyses if there is any
difference in the median values of three or more independent samples. The data values are ranked in an
increasing order, and the rank sums calculated followed by calculation of the test statistic.
Jonckheere test
In contrast to Kruskal–Wallis test, in Jonckheere test, there is an a priori ordering that gives it a more
statistical power than the Kruskal–Wallis test.[14]
Friedman test
The Friedman test is a non-parametric test for testing the difference between several related samples.
The Friedman test is an alternative for repeated measures ANOVAs which is used when the same
parameter has been measured under different conditions on the same subjects.[13]
Tests to analyse the categorical data Chi-square test, Fischer's exact test and McNemar's test are used
to analyse the categorical or nominal variables. The Chi-square test compares the frequencies and tests
whether the observed data differ significantly from that of the expected data if there were no
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 12/14
2/23/2019 Basic statistical tools in research and data analysis
differences between groups (i.e., the null hypothesis). It is calculated by the sum of the squared
difference between observed (O) and the expected (E) data (or the deviation, d) divided by the expected
data by the following formula:
A Yates correction factor is used when the sample size is small. Fischer's exact test is used to determine
if there are non-random associations between two categorical variables. It does not assume random
sampling, and instead of referring a calculated statistic to a sampling distribution, it calculates an exact
probability. McNemar's test is used for paired nominal data. It is applied to 2 × 2 table with paired-
dependent samples. It is used to determine whether the row and column frequencies are equal (that is,
whether there is ‘marginal homogeneity’). The null hypothesis is that the paired proportions are equal.
The Mantel-Haenszel Chi-square test is a multivariate test as it analyses multiple grouping variables. It
stratifies according to the nominated confounding variables and identifies any that affects the primary
outcome variable. If the outcome variable is dichotomous, then logistic regression is used.
There are a number of web resources which are related to statistical power analyses. A few are:
SUMMARY
It is important that a researcher knows the concepts of the basic statistical methods used for conduct of
a research study. This will help to conduct an appropriately well-designed study leading to valid and
reliable results. Inappropriate use of statistical techniques may lead to faulty conclusions, inducing
errors and undermining the significance of the article. Bad statistics may lead to bad research, and bad
research may lead to unethical practice. Hence, an adequate knowledge of statistics and the appropriate
use of statistical tests are important. An appropriate knowledge about the basic statistical methods will
go a long way in improving the research designs and producing quality medical research which can be
utilised for formulating the evidence-based guidelines.
Conflicts of interest
There are no conflicts of interest.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 13/14
2/23/2019 Basic statistical tools in research and data analysis
REFERENCES
1. Winters R, Winters A, Amedee RG. Statistics: A brief overview. Ochsner J. 2010;10:213–6.
[PMC free article] [PubMed]
3. Kaur SP. Variables in research. Indian J Res Rep Med Sci. 2013;4:36–8.
4. Satake EB. Statistical Methods and Reasoning for the Clinical Sciences Evidence-Based Practice. Ist
ed. San Diego: Plural Publishing, Inc; 2015. pp. 1–19.
5. Wilder RT, Flick RP, Sprung J, Katusic SK, Barbaresi WJ, Mickelson C, et al. Early exposure to
anesthesia and learning disabilities in a population-based birth cohort. Anesthesiology. 2009;110:796–
804. [PMC free article] [PubMed]
7. Myles PS, Gin T. Statistical Methods for Anaesthesia and Intensive Care. Ist ed. Oxford:
Butterworth Heinemann; 2000. pp. 8–10.
8. Binu VS, Mayya SS, Dhar M. Some basic aspects of statistical methods and sample size
determination in health science research. Ayu. 2014;35:119–23. [PMC free article] [PubMed]
9. Nickerson RS. Null hypothesis significance testing: A review of an old and continuing controversy.
Psychol Methods. 2000;5:241–301. [PubMed]
10. Mushtaq K, Ali Z, Shah NF, Syed S, Naqash I, Ramzan AU. A randomized controlled study to
compare the effectiveness of intravenous dexmedetomidine with placebo to attenuate the hemodynamic
and neuroendocrine responses to fixation of skull pin head holder for craniotomy. North J ISA.
2016;1:16–23.
11. Bajwa SJ. Basics, common errors and essentials of statistical tools and techniques in anesthesiology
research. J Anaesthesiol Clin Pharmacol. 2015;31:547–53. [PMC free article] [PubMed]
12. Das S, Mitra K, Mandal M. Sample size calculation: Basic principles. Indian J Anaesth.
2016;60:652–6. [PMC free article] [PubMed]
13. Altman DG, Bland JM. Parametric v non-parametric methods for data analysis. BMJ.
2009;338:a3167. [PubMed]
14. Bewick V, Cheek L, Ball J. Statistics review 10: Further nonparametric methods. Crit Care.
2004;8:196–9. [PMC free article] [PubMed]
15. Nahm FS. Nonparametric statistical tests for the continuous data: The basic concept and the
practical use. Korean J Anesthesiol. 2016;69:8–14. [PMC free article] [PubMed]
Articles from Indian Journal of Anaesthesia are provided here courtesy of Wolters Kluwer -- Medknow
Publications
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5037948/ 14/14