Module 5 Research Methodology (3)
Module 5 Research Methodology (3)
1. Large-Scale Data Handling: Computers enable the collection, storage, and analysis of
vast amounts of data, such as census data, survey responses, and social media
interactions, allowing researchers to uncover patterns and trends that would be impossible
to detect manually.
2. Quantitative Analysis: Statistical software like SPSS, Stata, and R allows sociologists to
perform complex statistical analyses, including regression and multivariate analysis,
enabling precise and reliable hypothesis testing.
3. Qualitative Analysis: Software such as NVivo and Atlas.ti aids in coding and analyzing
qualitative data from interviews, focus groups, and ethnographic notes, helping
researchers identify themes and patterns across large volumes of textual data.
5. Big Data and Social Media Analysis: The analysis of big data and social media
interactions is facilitated by tools like Python and specialized APIs, enabling researchers
to scrape, clean, and analyze data from online platforms and digital footprints.
6. Collaborative Research: Online platforms and tools such as cloud storage and
collaborative documents enable global collaboration among researchers, supporting real-
time communication and data sharing, thus enhancing the efficiency of research projects.
DATA ANALYSIS
Data analysis is a comprehensive method of inspecting, cleansing, transforming, and modeling
data to discover useful information, draw conclusions, and support decision-making. It is a
multifaceted process involving various techniques and methodologies to interpret data from
various sources in different formats, both structured and unstructured.
The first step in the data analysis process is to define the objectives and formulate clear, specific
questions that your analysis aims to answer. This step is crucial as it sets the direction for the
entire process. It involves understanding the problem or situation at hand, identifying the data
needed to address it, and defining the metrics or indicators to measure the outcomes.
Data cleaning, also known as data cleansing, is a critical step in the data analysis process. It
involves checking the data for errors and inconsistencies, and correcting or removing them. This
step ensures the quality and reliability of the data, which is crucial for obtaining accurate and
meaningful results from the analysis.
Once the data is cleaned, it's time for the actual analysis. This involves applying statistical or
mathematical techniques to the data to discover patterns, relationships, or trends. There are
various tools and software available for this purpose, such as Python, R, Excel, and specialized
software like SPSS and SAS.
After the data is analyzed, the next step is to interpret the results and visualize them in a way that
is easy to understand. This could involve creating charts, graphs, or other visual representations
of the data. Data visualization helps to make complex data more understandable and provides a
clear picture of the findings.
The final step in the data analysis process is data storytelling. This involves presenting the
findings of the analysis in a narrative form that is engaging and easy to understand. Data
storytelling is crucial for communicating the results to non-technical audiences and for making
data-driven decisions.
Diagnostic analysis- Diagnostic analysis goes a step further than descriptive analysis by
determining why something happened. It involves more detailed data exploration and comparing
different data sets to understand the cause of a particular outcome. For instance, if a company's
sales dropped in a particular month, diagnostic analysis could be used to find out why.
Predictive analysis- Predictive analysis uses statistical models and forecasting techniques to
understand the future. It involves using data from the past to predict what could happen in the
future. This type of analysis is often used in risk assessment, marketing, and sales forecasting.
For example, a company might use predictive analysis to forecast the next quarter's sales based
on historical data.
Prescriptive analysis- Prescriptive analysis is the most advanced type of data analysis. It not
only predicts future outcomes but also suggests actions to benefit from these predictions. It uses
sophisticated tools and technologies like machine learning and artificial intelligence to
recommend decisions. For example, a prescriptive analysis might suggest the best marketing
strategies to increase future sales.
MEAN
Mean is the average of the given numbers and is calculated by dividing the sum of given
numbers by the total number of numbers.
TYPES OF MEAN
Arithmetic Mean- When you add up all the values and divide by the number of values it is
called Arithmetic Mean. To calculate, just add up all the given numbers then divide by how
many numbers are given.
Geometric Mean - It is the average value or mean which signifies the central tendency of the set
of numbers by finding the product of their values.
Harmonic Mean- It is defined as the reciprocal of the average of the reciprocals of the data
values.. It is based on all the observations, and it is rigidly defined.
MERITS OF MEAN
1. Simplicity and Ease of Calculation
MEDIAN
In Mathematics, the median is defined as the middle value of a sorted list of numbers. The
middle number is found by ordering the numbers. The numbers are ordered in ascending order.
Once the numbers are ordered, the middle number is called the median of the given data set.
It is easy to find the median for the dataset, that has an odd number of observations.
Eg. Median of 2, 5, 8 is 5
If the dataset is even, then the mean value or average for the middle two numbers is called the
median of the given data set.
MERITS OF MEDIAN
1. Simple to understand and calculate
DEMERITS OF MEDIAN
1. Less efficient for large datasets
MODE
A mode is defined as the value that has a higher frequency in a given set of values. It is the value
that appears the most number of times.
Example: In the given set of data: 2, 4, 5, 5, 6, 7, the mode of the data set is 5 since it has
appeared in the set twice.
CHARACTERSTICS OF MODE
1. Represents the most frequently occurring value in a dataset.
MERITS
1. Easy to identify and calculate.
3. Resistant to outliers.
DEMERITS
May not be representative of the entire dataset, especially in datasets with multiple modes
or when the mode is significantly different from other values.
Not sensitive to the actual values of data points, only their frequencies.
Limited applicability to continuous data, as it may not accurately represent the central
tendency.
In some cases, there may be no mode or multiple modes, making it less informative for
describing the dataset.
DATA INTERPRETATION
Data Interpretation refers to the process of using diverse analytical methods for making sense of
a collection of data that has been processed. The collected data may be present in various forms
like bar graphs, line charts, histograms, pie charts, tabular forms etc and hence it needs to be
interpreted to summarise the information.
Data Interpretation is the process of understanding, organising, and interpreting the given data,
for making sense of and getting a meaningful conclusion. The basic concept of data
interpretation is to review the collected data by means of analytical methods and arrive at
relevant conclusions. There are two methods to interpret the data:
1. Qualitative method – This method is used to analyse qualitative data or categorical data.
The qualitative data interpretation used texts instead of numbers or patterns to represent
the data. Nominal and ordinal data are the two types of qualitative data. Ordinal data
interpretation is much easier than nominal data interpretation.
2. Quantitative method -This method is used to analyse quantitative data or numerical data.
Quantitative data interpretation uses numbers instead of texts to represent the data. The
types of quantitative data interpretation are discrete and continuous data. The quantitative
method of data interpretation requires statistical methods and techniques like mean,
median, standard deviation, etc. to interpret the data.
Pie Chart – the circular graph used to represent the percentage of a variable is called a pie
chart. The pie charts represent numbers as proportions or percentages. Some types of pie
charts are simple pie charts, doughnut pie charts, and 3D pie charts.
Tables – statistical data are represented by tables. The data are placed in rows and
columns. Types of tables include simple tables and complex tables.
Line Graph – the charts or graphs that show information in a series of points are included
in the line graphs. Line charts are very good to visualise continuous data or sequence of
values. Some of the types of line graphs are simple line graphs, stacked line graphs etc.
DATA INFERENCING
Data inferencing refers to the process of drawing conclusions or making predictions based on
data analysis. It involves using statistical and computational methods to derive insights from
data, which can then inform decision-making, policy formulation, and scientific research
Graphs, charts, tables, and summary Statistical tests, estimation, hypothesis testing,
Methods Used statistics. predictive modeling.
Comparing data with benchmarks, Applying models to sample data to infer about
Contextualization historical data, or theoretical expectations. the population or predict future events.
Describing trends in sales data over the Predicting future sales based on historical
Examples past year. data.
GENERALIZATION
Generalization refers to the cognitive process of forming broad concepts or conclusions from
specific instances or observations. It involves identifying patterns or commonalities among
particular examples and extending these patterns to apply to a wider set of circumstances.
SCHEDULE
The schedule is also one of the methods of data collection. It will have a set of statements,
questions and space given to note down the answers. The Schedule method of data collection can
be utilised irrespective of the respondent’s literacy. It can be used when the respondents are
literate and can be used even when the respondents are illiterate.
ESSENTIALS
1. Timeline for Each Research Phase: A schedule in research methodology should outline
specific timeframes for each phase of the research process, including planning, data
collection, analysis, and reporting. This ensures that the research progresses
systematically and is completed within a reasonable timeframe.
PYQ
Q. Explain the meaning of data interpretation(2023) 1 marks