Professional Documents
Culture Documents
Research Data Interpretation
Research Data Interpretation
INTERPRETATION
BY
PROF A. BALASUBRAMANIAN
CENTRE FOR ADVANCED STUDIES IN EARTH
SCEINCE
UNIVERSITY OF MYSORE, INDIA
‘All meanings, we know, depend on
the key of interpretation.’
-George Eliot
2
Principles of Analysis and Interpretation
Data, as used in behavioral research, means
research results from which inferences are drawn:
usually numerical results, like scores of tests and
statistics such as means, percentages, and
correlation coefficients.
Analysis means the categorizing, ordering,
manipulating, and summarizing of data to obtain
answers to research questions.
Interpretation takes the results of analysis, makes
inferences pertinent to the research relations
studied, and draws conclusions about these
relations.
Methods of data interpretation
Direct visual observations of raw data
After organizing the data in tables
After making Graphical representations
After calculations using numerical /
statistical methods
After mathematical modelling
DATA
Data is known to be crude information
and not knowledge by itself.
The sequence from data to knowledge
is:
from Data to Information,
from Information to Facts, and finally,
from Facts to Knowledge.
DATA
Data becomes information, when it becomes
relevant to your decision problem.
Information becomes fact, when the data can
support it.
Facts are what the data reveals.
However the decisive instrumental (i.e., applied)
knowledge is expressed together with
some statistical degree of confidence.
Fact becomes knowledge, when it is
used in the successful completion
of a decision process.
massive amount of facts are
integrated as knowledge.
Usefulness and utility of research findings lie in proper
interpretation.
Interpretation is a basic component of research.
After collecting and analyzing the data, the researcher
has to accomplish the task of drawing inferences
followed by report writing.
This has to be done very carefully, otherwise mis
conclusions may be drawn and the whole purpose of
doing research may get vitiated.
It is only through interpretation that the researcher
can expose relations and processes that underlie his
findings.
Meaning of Interpretation
Interpretation refers to the task of drawing inferences
from the collected facts after an analytical and or
experimental study.
In fact, it is a search for broader meaning of research
findings.
The task of interpretation has two major aspects viz.,
the effort to establish continuity in research through
linking the results of a given study with those of
another, and the establishment of some explanation
concepts.
“In one sense, interpretation is concerned with
relationships within the collected data, partially
overlapping analysis.
Interpretation also extends beyond the data of
the study to inch the results of other research,
theory and hypotheses.”
Interpenetration is the device
Thus, interpenetration is the device through
which the factors that seem to explain what
has been observed by researcher in the
course of the study can be better understood
and it also provides a theoretical conception
which can serve as a guide for further
researches.
Why Interpretation?
Interpretation is essential for the simple
reason that the usefulness and utility of
research findings lie in proper
interpretation.
It is being considered a basic component of
research process because of the following
reasons:
Through interpretation
It is through interpretation that the researcher
can well understand the abstract principle that
works beneath his findings.
Through this he can link up his findings with
those of other studies, having the same abstract
principle, and thereby can predict about the
concrete world of events. Fresh inquiries can test
these predictions later on.
This way the continuity in research can be
maintained.
Interpretation leads to establishment
Interpretation leads to the establishment of
explanatory concepts that can serve as a guide for
future research studies;
it opens new avenues of intellectual adventure
and stimulates the quest for more knowledge.
Researcher can better appreciate only through
interpretation why his findings are what they are
and can make others to understand the real
significance of his research findings.
Interpretation of the findings
The interpretation of the findings of exploratory
research study often results into hypotheses for
experimental research and as such interpretation
is involved in the transition from exploratory to
experimental research.
Since an exploratory study does not have a
hypothesis to start with, the findings of such a
study have to be interpreted on a post factum
basis in which case the interpretation is
technically described as ‘post factum’
interpretation.
Technique of Interpretation
The task of interpretation is not an easy job,
rather it requires a great skill and dexterity on the
part of researcher.
Interpretation is an art that one learns through
practice and experience.
The researcher may, at times, seek the guidance
from experts for accomplishing the task of
interpretation.
The technique of interpretation
The technique of interpretation often involves the
following steps:
1. Researcher must give reasonable explanations
of the relations which he/she has found and
he/she must interpret the lines of relationship
in terms of the underlying processes and must
try to find out the thread of uniformity that lies
under the surface layer of his diversified
research findings.
In fact, this is the technique of how generalization
should be done and concepts be formulated.
2. Extraneous information, if collected
during the study, must be considered
while interpreting the final results of
research study, for it may prove to be a
key factor in understanding the problem
under consideration.
3. It is advisable, before embarking upon final interpretation,
to , consult someone having insight into the study and who
is frank and honest and will not hesitate to point out
omissions and errors in logical argumentation. Such a
consultation will result in correct interpretation and, thus,
will enhance the utility of research results.
4. Researcher must accomplish the task of interpretation only
after considering all relevant factors affecting the problem
to avoid false generalization. He /she must be in no hurry
while interpreting results, for quite often the conclusions,
which appear to be all right at the beginning, may not at all
be accurate.
Precautions in Interpretation
One should always remember that even if
the data are properly collected and analyzed,
wrong interpretation would lead to
inaccurate conclusions.
It is, therefore, absolutely essential that the
task of , interpretation be accomplished
with patience in an impartial manner and
also in correct perspective.
For correct interpretation
Researcher must pay attention to the following points
for correct interpretation:
(i) At the outset, researcher must invariably satisfy
himself that
(a) the data are appropriate, trustworthy and adequate
for drawing inferences;
(b) the data reflect good homogeneity; and that
(c) proper analysis has been done through statistical
methods.
(ii) The researcher must remain cautious about the
errors that can possibly arise in the process of
interpreting results.
Errors can arise due to false
generalization
Errors can arise due to false generalization and/or due
to wrong interpretation of statistical measures, such as
the application of findings beyond the range of
observations, identification of correlation with
causation . and the like.
Another major pitfall is the tendency to affirm that
definite relationships exist on the basis of
confirmation of particular hypotheses.
Researcher must remain vigilant
In fact, the positive test results accepting the
hypothesis must be interpreted as “being in
accord” with the hypothesis, rather than as
“confirming the validity of the hypothesis”.
The researcher must remain vigilant about all
such things so that false generalization may not
take place.
He/she should be well equipped with and must
know the correct use of statistical measures for
drawing inferences concerning his study.
Researcher must always keep in view that the
task of interpretation is very much intertwined
with analysis and cannot be distinctly separated.
As such he must take the task of interpretation as
a special aspect of analysis and accordingly must
take all those precautions that one usually
observes while going through the process of
analysis viz., precautions concerning the
reliability of data, computational checks,
validation and comparison of results.
Researcher must never lose sight of the fact that his
task is not only to make sensitive observations of
relevant occurrences, but also to identify and
disengage the factors that are initially hidden to the
eye.
This will enable him to do his job of interpretation on
proper lines.
Broad generalization should be avoided as most
research is not amen- able to it because the coverage
may be restricted to a particular time, a particular area
and particular conditions.
Such restrictions, if any, must invariably be specified
and the results must be framed within their limits.
The researcher must remember that “ideally in the
course of a research study, there should be constant
interaction between initial hypothesis, empirical
observation and theoretical conceptions.
It is exactly in this area of interaction between
theoretical orientation and empirical observation that
opportunities for originality and creativity lie.”
Researcher must pay special attention to this aspect
while engaged in the task of interpretation.
Data Interpretation Methods
Data interpretation may be the most important key in
proving or disproving your hypothesis.
It is important to select the proper statistical tool to
make useful interpretation of your data.
If you pick an improper data analysis method, your
results may be suspect and lack credibility.
Visually scanning the data
Before doing any statistical analyses of
the data you have collected, look closely
at the data to determine the best
method of organizing it .
By visually scanning the data and
reorganizing it, you may be able to spot
trends or other anomalies that may help
you in your analysis of the data.
STATISTICS
Statistics is a science assisting you to
make decisions under
uncertainties (based on some numerical
and measurable scales).
Decision making process must be based on
data neither on personal opinion nor on any
belief.
What is Statistical Data Analysis?
Data are not information! To
determine what statistical data
analysis is, one must first define
statistics.
Statistics is a set of methods that
are used to collect, analyze, present,
and interpret data.
Statistical methods
Statistical methods are used in a wide
variety of occupations and help people
identify, study, and solve many complex
problems.
In the business and economic world,
these methods enable decision makers
and managers to make informed and
better decisions about uncertain
situations.
Review-Statistics:
“We can think of statistics as a group of
computational procedures that allow us to find
meaning in numerical data”.
Descriptive statistics provide a description of
what the data look like.
They provide a means to describe the points of
central tendency (mean, mode, median, etc.) and
dispersion (standard deviation, variance, inter-
quartile range, etc.).
Inferential statistics allow the
researcher to make
inferences about populations from smaller samples of the
population.
Statistics of the sample are used to estimate parameters of
the population.
A parameter is a constant value representative of the
population (such as population mean and standard
deviation) while a statistic is any calculation performed on
the sample being tested.
Inferential statistics also allow the researcher to test their
research hypotheses.
Some measures used in inferential statistics include the
standard error of the mean, estimators, and the p-value.
The way that the data is interpreted can have
varying effects on your conclusions.
Absolute honesty in recording and interpreting
data is required to maintain the credibility of
research.
All of the conditions of a situation should be
considered and that we make inferences in strict
accordance with the data obtained.
Using statistics to determine relationships is
paramount to the success of good research.
Using tools such as ANOVA, correlations, Fisher
Exact Tests, regression, etc. can predict whether or
not your research hypothesis is satisfied.
But, REMEMBER to select your p-value before you
begin your research project.
Doing this will add credibility to your research.
One other important point to remember when
doing data analysis is to use parametric statistics
instead of nonparametric statistics whenever
possible.
Remember that parametric statistics relies on the
assumptions of normality, which gives greater
power than nonparametric testing.
Both parametric and nonparametric statistical
tests are used for interpretation.
Interpreting Qualitative Data:
Qualitative data interpretation tends to be more
subjective in nature and many times can be influenced
by the researcher’s biases.
Effort must be put into the data collection process to
eliminate bias including collecting more than one kind
of data, get many different kinds of perspectives on the
events being studied, purposely look for contradicting
information, and acknowledging your biases that
relate to your research report.
Qualitative data analysis
Qualitative data analysis is time consuming and
complex because a lot of data can be created that is
both useful and not useful.
There is no “correct way” to analyze qualitative data.
Efforts can be made to make your data presentation
and interpretation more credible and less biased by
using the above methods.
Statistics consists of
Statistics consists of the principles and methods for
Designing studies
Collecting data
Presenting and analysing data
Interpreting the results
Statistics has been described as
Turning data into information
Data-based decision making.
Data interpretation:
Uncovering and explaining trends in the data.
The analyzed data can then be interpreted and
explained.
In general, when scientists interpret data, they
attempt to explain the patterns and trends
uncovered through analysis, bringing all of their
background knowledge, experience, and skills to
bear on the question and relating their data to
existing scientific ideas.
Given the personal nature of the
knowledge they draw upon, this step can
be subjective, but that subjectivity is
scrutinized through the peer
review process.
Data collection is the systematic
recording
Data collection is the systematic recording of
information;
data analysis involves working to uncover
patterns and trends in datasets;
data interpretation involves explaining those
patterns and trends.
Scientists interpret data based on
their background
Scientists interpret data based on their
background knowledge and experience; thus,
different scientists can interpret the same data in
different ways.
By publishing their data and the techniques they
used to analyze and interpret those data, scientists
give the community the opportunity to both
review the data and use them in future research.
Data Analysis
Data Analysis is the process of systematically
applying statistical and/or logical techniques to
describe and illustrate, condense and recap, and
evaluate data.
Various analytic procedures “provide a way of drawing
inductive inferences from data and distinguishing the
signal (the phenomenon of interest) from the noise
(statistical fluctuations) present in the data”..
While data analysis in qualitative research
can include statistical procedures, many
times analysis becomes an ongoing iterative
process where data is continuously collected
and analyzed almost simultaneously.
Indeed, researchers generally analyze for
patterns in observations through the entire
data collection phase.
The form of the analysis is determined by
the specific qualitative approach taken (field
study, ethnography content analysis, oral
history, biography, etc) and
the form of the data (field notes,
documents, audiotape, videotape).
An essential component of ensuring data integrity
is the accurate and appropriate analysis of research
findings.
Improper statistical analyses distort scientific
findings, mislead casual readers, and may
negatively influence the public perception of
research.
Integrity issues are just as relevant to analysis of
non-statistical data as well.
Considerations/issues in data
analysis
There are a number of issues that researchers should be cognizant of with respect to data
analysis. These include:
Having the necessary skills to analyze
Concurrently selecting data collection methods and appropriate analysis
Drawing unbiased inference
Inappropriate subgroup analysis
Following acceptable norms for disciplines
Determining statistical significance
Lack of clearly defined and objective outcome measurements
Providing honest and accurate analysis
Manner of presenting data
Environmental/contextual issues
Data recording method
Partitioning ‘text’ when analyzing qualitative data
Training of staff conducting analyses
Reliability and Validity
Extent of analysis
Summarizing data
Tables
Simplest way to summarize data
Data are presented as absolute numbers or percentages
Charts and graphs
Visual representation of data
Data are presented as absolute numbers or percentages
Basic guidance when summarizing
data
Ensure graphic has a title
Label the components of your graphic
Indicate source of data with date
Provide number of observations (n=xx) as a reference
point
Add footnote if more information is needed
Set of categories with numerical counts
10%
1st Qtr
2nd Qtr
3rd Qtr
23% 59% 4th Qtr
N=150
Pie Chart
Use it to…
convey approximate proportional relationships
(relative amounts) at a point in time
compare part of a whole at a given point in time
emphasize a small proportion of parts
Do NOT use it…
For exact comparisons of values, because estimating
angles is difficult for people
For rank data: Use column/bar charts in this case; use
multiple column/bar charts for grouped data
If proportions vary greatly; do not use multiple pies to
compare corresponding parts
Caution!
4
Site 1
3
Site 2
2 Site 3
0
Quarter 1 Quarter 2 Quarter 3 Quarter 4
Percentage of new enrollees tested for HIV at each site, by
quarter
6
% o f new enrollees tested for
5
4
3
HIV
2
Site 1
1 Site 2
0 Site 3
Quarter 1 Quarter 2 Quarter 3 Quarter 4
Q1 Jan–Mar Q2 Apr–June Q3 July–Sept Q4 Oct–Dec
Months
Has the program met its goal?
Percentage of new enrollees tested for HIV at each site, by
quarter
60%
% of new enrollees tested
50%
40%
for HIV
30% Site 1
20% Site 2
Site 3
10%
Target
0%
Quarter 1 Quarter 2 Quarter 3 Quarter 4
Stacked bar chart
Represent components of whole & compare wholes
Number of Months Female and Male Patients Have Been
Enrolled in HIV Care, by Age Group
Females 4 10
0-14 years
15+ years
Males 3 6
0 5 10 15
Number of months patients have been enrolled in HIV care
Line graph
Displays trends over time
Number of Clinicians Working in Each Clinic During Years 1–4*
5
Number of clinicians
4
Clinic 1
3
Clinic 2
2 Clinic 3
1
0
Year 1 Year 2 Year 3 Year 4
Number of Clinicians Working in Each Clinic During Years 1-4*
Line graph 6
5
Number of clinicians
4
Clinic 1
3
Clinic 2
2 Clinic 3
0
Year 1
Y1 1995 19962
Y2Year 19973
Y3Year Y4 1998
Year 4
Interpreting data
Adding meaning to information by making
connections and comparisons and exploring causes
and consequences
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Interpretation – relevance of
finding
Adding meaning to information by making
connections and comparisons and exploring causes
and consequences
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Interpretation – relevance of
finding
Does the indicator meet the target?
How far from the target is it?
How does it compare (to other time periods, other
facilities)?
Are there any extreme highs and lows in the data?
Interpretation – possible causes?
• Supplement with expert opinion
• Others with knowledge of the program or target
population
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Interpretation – consider other
data
Use routine service data to clarify questions
• Calculate nurse-to-client ratio, review
commodities data against client load, etc.
Use other data sources
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Interpretation – other data
sources
Situation analyses
Demographic and health surveys
Performance improvement data
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Interpretation – conduct further
research
Data gap conduct further research
Methodology depends on questions being asked
and resources available
Conduct
Relevance Reasons for Consider
further
of finding finding other data
research
Data Interpretation
Answer these four questions
What is important in the data?
Why is it important?
What can be learned from it?
So what?
Remember
Interpretation depends on the perspective of the
researcher.
Why?
Interpretation
One technique for data interpretation (Wolcott)
Extend the analysis by raising questions
Connect findings to personal experiences
Seek the advice of “critical” friends.
Contextualize findings in the research
Converging evidence?
Turn to theory
Frequencies and Continuous Measures
Quantitative data come in two general forms:
frequencies and continuous measures.
f={(x,y); where x is a member of the set X, and y is
either 1 or 0 depending on x’s possessing or not
possessing M}
f={(x,y); x is an object, and y= any numeral}
Rules of Categorization
The first setup in any analysis is categorization.
The five rules of categorization are as follows:
1.Categories are set up according to the research
problem and purpose.
2.The categories are exhaustive.
3.The categories are mutually exclusive and
independent.
4.Each category (variable) is derived from one
classification principle.
5.Any categorization scheme must be on one level
of discourse
Kinds of Statistical Analysis
Frequency Distributions
Graphs and Graphing
Measures of Central Tendency and Variability
Measures of Relations
Analysis of Differences
Analysis of Variance and Related Methods
Profile Analysis
Multivariate Analysis
Graphs and Graphing
A graph is a two-dimension representation of a
relation or relations.
Interaction means that the relation of an independent
variable to a dependent variable differs in different
groups or at different levels of another independent
variable..
Frequency Distributions
Although frequency distributions are used primarily
for descriptive purposes, they can also be used for
other research purposes.
Observed distributions can also be compared to
theoretical distributions (normal distributions).
Measures of Central Tendency and
Variability
Mean, median, mode
Standard deviation, range
Measures of Relations
Ideally, any analysis of research data should include
both kinds of indices: measures of the significance of a
relation and measures of the magnitude of the relation.
Analysis of Differences
1.it is by no means confined to the differences between
measures of central tendency.
2.All analyses of differences are intended for the
purpose of studying relation. Conversely, the greater
the differences the higher the correlation, all other
things being equal.
Analysis of Variance and Related Methods
A method of identifying, breaking down, and testing
for statistical significance variances that come from
different sources of variation.
That is, a dependent variable has a total amount of
variance, some of which is due to the experimental
treatment, some to error, and some to other causes.
Profile Analysis
Profile analysis is basically the assessment of the
similarities of the profiles of individuals or groups.
A profile is a set of different measures of an individual
or group, each of which is expressed in the same unit
of measure.
Multivariate Analysis
Multiple regression
Canonical correlation
Discriminant analysis
Factor analysis
Cluster analysis
Path analysis
Analysis of covariance structures
Log-linear models
Indices
Index can be defined in two related ways:
1.An index is an observable phenomenon that is
substituted for a less-observable phenomenon. For
example, test scores indicate achievement levels,
verbal aptitudes, degrees of anxiety, and so on.
2.An index is a number that is a composite of two or
more numbers. For example, all sums and averages,
coefficients of correlation.
Indices
Indices are most important in research because they
simplify comparisons.
The percentage is a good example.
Percentages transform raw numbers into comparable
form.
Indices generally take the form of quotients: ratios and
proportions.
Social Indicators
Indicators, although closely related to indices—
indeed, they are frequently indices as defined
above—form a special class of variables.
Variables like income, life expectancy, fertility,
quality of life, educational level (of people), and
environment can be called social indicators. Social
indicators are both variables and statistics.
Unfortunately, it is difficult to define “social indicators.”
In this book we are interested in social indicators as a class
of sociological and psychological variables that in the
future may be useful in developing and testing scientific
theories of the relations among social and psychological
phenomena.
The Interpretation of Research Data
Adequacy of Research Design, Methodology,
Measurement, and Analysis
Negative and Inconclusive Results
Unhypothesized Relations and Unanticipated
Findings
Proof, Probability, and Interpretation
Adequacy of Research Design, Methodology,
Measurement, and Analysis
Most important, the design, methods of observation,
measurement, and statistical analysis must all be
appropriate to the research problem.
Negative and Inconclusive Results
When results are positive, when the data support
the hypotheses, one interprets the data along the
lines of the theory and the reasoning behind the
hypotheses.
If we can repeat the feat, the n the evidence of
adequacy is even more convincing.
If we can be fairly sure that the methodology, the
measurement, and the analysis are adequate, then
negative results can be definite contributions to
scientific advancement.
Unhypothesized Relations and
Unanticipated Findings
The unpredicted relation may be an important key
to a deeper understanding of the theory.
For example, positive reinforcement strengthens
response tendencies.
Unpredicted and unexpected findings must be
treated with more suspicion than predicted and
expected findings.
Before being accepted, they should be
substantiated in independent research in which
they are specially predicted and tested.
Proof, Probability, and Interpretation
Let us flatly assert that nothing can be “proved”
scientifically.
All one can do is to bring evidence to bear that
such-and such a proposition is true.
Proof is a deductive matter.
Experimental methods of inquiry are not methods
of proof, they are controlled methods of bringing
evidence to bear on the probable truth or falsity of
relational propositions.
In short, no single scientific investigation ever
proves anything.
Thus the interpretation of the analysis of research
data should never use the word proof.
Effective Data Analysis
Effective data analysis involves
keeping your eye on the main game
managing your data
engaging in the actual process of quantitative and / or
qualitative analysis
presenting your data
drawing meaningful and logical conclusions
The Big Picture
Analysis should be approached as a critical, reflective,
and iterative process that cycles between data and an
overarching research framework that keeps the big
picture in mind
116
Managing Data
Regardless of data type, managing your data
involves
familiarizing yourself with appropriate software
developing a data management system
systematically organizing and screening your data
entering the data into a program
and finally ‘cleaning’ your data
117
Statistics
Being able to do statistics no longer means being able
to work with formula
It’s much more important for researchers to be familiar
with the language and logic of statistics, and be
competent in the use of statistical software
Data Types
Different data types demand discrete treatment, so it’s
important to be able to distinguish variables by
cause and effect (dependent or independent)
measurement scales (nominal, ordinal, interval, and
ratio)
Descriptive Statistics
Descriptive statistics are used to summarize the basic
feature of a data set through
measures of central tendency (mean, mode, and
median)
dispersion (range, quartiles, variance, and standard
deviation)
distribution (skewness and kurtosis)
Inferential Statistics
Inferential statistics allow researchers to assess
their ability to draw conclusions that extent
beyond the immediate data, e.g.
if a sample represents the population
if there are differences between two or more groups
if there are changes over time
if there is a relationship between two or more variables
Selecting Statistical Tests
Selecting the right statistical test relies on
knowing the nature of your variables
their scale of measurement
their distribution shape
types of question you want to ask
Presenting Quantitative Data
Presenting quantitative data often involves the
production of graphs and tables
These need to be
1. selectively generated so that they make relevant
arguments
2. informative yet simple, so that they aid reader’s
understanding
Qualitative Data Analysis (QDA)
In qualitative data analysis there is a common reliance
on words and images to draw out rich meaning
But there is an amazing array of perspectives and
techniques for conducting an investigation
The QDA Process
Qualitative data analysis creates new
understandings by exploring and interpreting
complex data from sources without the aid of
quantification
Data source include
interviews
group discussions
observation
journals
archival documents, etc
Uncovering and Discovering Themes
The methods and logic of qualitative data analysis
involve uncovering and discovering themes that run
through raw data, and interpreting the implication of
those themes for research questions
126
More on the QDA Process
Qualitative data analysis generally involves
moving through cycles of inductive and deductive
reasoning
thematic exploration (based on words, concepts, literary
devises, and nonverbal cues)
exploration of the interconnections among themes.
Specialist QDA Strategies
There are a number of paradigm and discipline
based strategies for qualitative data analysis
including
content analysis
discourse analysis
narrative analysis
conversation analysis
semiotics
hermeneutics
grounded theory
128
Presenting Qualitative Data
Effective presentation of qualitative data can be a real
challenge
You’ll need to have a clear storyline, and selectively use
your words and/or images to give weight to your story
Drawing Conclusions
Your findings and conclusions need to flow from
analysis and show clear relevance to your overall
project
Findings should be considered in light of
significance
current research literature
limitations of the study
your questions, aims, objectives, and theory
Add Interpretation of Analysis
of Data
Include your interpretation:
What does the data MEAN with regards to that theme?
The “So what?” of the theme and/or data.
WHY DO WE ANALYZE DATA
The purpose of analysing data is to obtain usable and
useful information.
The analysis, irrespective of whether the data is
qualitative or quantitative, may:
•describe and summarise the data
•identify relationships between variables
•compare variables
•identify the difference between variables
•forecast outcomes
SCALES OF MEASUREMENT
Many people are confused about what type of analysis
to use on a set of data and the relevant forms of
pictorial presentation or data display.
The decision is based on the scale of measurement of
the data.
These scales are
nominal,
Ordinal and
numerical.
Nominal scale
A nominal scale is where:
The data can be classified into a non-numerical or
named categories and
The order in which these categories can be written or
asked is arbitrary.
Ordinal scale
An ordinal scale is where:
the data can be classified into non-numerical or
named categories -an inherent order exists among the
response categories.
Ordinal scales are seen in questions that call for
ratings of quality (for example, very good, good, fair,
poor, very poor) and agreement (for example, strongly
agree, agree, disagree, strongly disagree).
Numerical scale
A numerical scale is:
where numbers represent the possible response
categories
there is a natural ranking of the categorieszero on the
scale has meaningthere is a quantifiable difference
within categories and between consecutive categories.
QUALITATIVE ANALYSIS
"Data analysis is the process of bringing order,
structure and meaning to the mass of collected data. It
is a messy, ambiguous, time-consuming, creative, and
fascinating process.
It does not proceed in a linear fashion; it is not neat.
Qualitative data analysis is a search for general
statements about relationships among categories of
data."
Simple qualitative analysis
Unstructured -are not directed by a script. Rich but
not replicable.
•Structured -are tightly scripted, often like a
questionnaire.
Replicable but may lack richness.
•Semi-structured -guided by a script but interesting
issues can be explored in more depth.
Can provide a good balance between richness and
replicability.
Simple qualitative analysis
Recurring patterns or themes
–Emergent from data, dependent on observation
framework if used
•Categorizing data
–Categorization scheme may be emergent or pre-
specified
•Looking for critical incidents
–Helps to focus in on key events
TOOLS TO SUPPORT DATA ANALYSIS
•Spreadsheet –simple to use, basic graphs
•Statistical packages, e.g. SPSS
•Qualitative data analysis tools
–Categorization and theme-based analysis, e.g. N6
–Quantitative analysis of text-based data
Interpreting research results
Researchers should describe their results clearly, and
in a way that other researchers can compare them with
their own results. They should also analyse the results,
using appropriate statistical methods to try to
determine the probability that they may have been
chance findings, and may not be replicable in larger
studies. But this is not enough.
Results need to be interpreted in an objective and
critical way, before assessing their implications and
before drawing conclusions.
Interpretation of research results is not just a concern
for researchers.
Policymakers should also be aware of the possible
pitfalls in interpreting research results and should be
cautious in drawing conclusions for policy decisions.
Interpreting descriptive statistics
The mean or average is only meaningful if the data fall into
a normal distribution curve, that is, they are evenly
distributed around the mean.
The mean or average, by itself, has a limited value.
There is an anecdote about a man having one foot on ice
and the other in boiling water; statistically speaking, on
average, he is pretty comfortable.
The range of the data, and their distribution (expressed in
the standard deviation) must be known.
It is sometimes more important to know the number or
percentage of subjects or values that are abnormal than to
know the mean
Representation v/s interpretation
Tables/charts/graphs
Diagrams
Maps/ atlases/ sketches
Profiles/ cross-sections
Plots
3-D views
Animations
Significance of Report Writing
Research report is considered a major component of
the research study for the research task remains
incomplete till the report has been presented and/or
written.
As a matter of fact even the most brilliant hypothesis,
highly well designed and conducted research study,
and the most striking generalizations and findings are
of little value unless they are effectively communicated
to others.
The purpose of research is not well served unless the
findings are made known to others.
Research results must invariably enter the general
store of knowledge.
All this explains the significance of writing research
report.
There are people who do not consider writing of report
as an integral part of the research process.
But the general opinion is in favour of treating the
presentation of research results or the writing of
report as part and parcel of the research project.
Writing of report is the last step in a research study
and requires a set of skills somewhat different from
those called for in respect of the earlier stages of
research.
This task should be accomplished by the researcher
with utmost care; he may seek the assistance and
guidance of experts for the purpose.
Different Steps in Writing Report
Research reports are the product of slow, painstaking,
accurate inductive work.
The usual steps involved in writing report are:
(a) logical analysis of the subject-matter;
(b) preparation of the final outline;
(c) preparation of the rough draft;
(d) rewriting and polishing;
(e) preparation of the final bibliography; and
(f) writing the final draft.
Though all these steps are self explanatory, yet a brief
mention of each one of these will be appropriate for
better under-standing.
Logical analysis of the subject matter.
It is the first step which is primarily concerned with
the development of a subject.
There are two ways in which to develop a subject—
(a) logically and
(b) chronologically.
The logical development is made on the basis of
mental connections and associations between one
thing and another by means of analysis.
Logical treatment often consists in developing the
material from the simple possible to the most complex
structures.
Chronological development is based on a connection
or sequence in time or occurrence.
The directions for doing or making something usually
follow the chronological order.
Preparation of the final outline. It is the next step in
writing the research report.
“Outlines are the framework upon which long written
works are constructed.
They are an aid to the logical organisation of the
material and a reminder of the points to be stressed in
the report.”
Preparation of the rough draft
This follows the logical analysis of the subject and the
preparation of the final outline.
Such a step is of utmost importance for the researcher now
sits to write down what he has done in the context of his
/her research study.
He/she will write down the procedure adopted by him/her
in collecting the material for his study along with various
limitations faced by him/her, the technique of analysis
adopted by him/her, the broad findings and
generalizations and the various suggestions he /she wants
to offer regarding the problem concerned.
Rewriting and polishing of the rough draft. This step
happens to be the most difficult part of all formal writing.
Usually this step requires more time than the writing
of the rough draft. The careful revision makes the
difference between a mediocre and a good piece of
writing.
While rewriting and polishing, one should check the
report for weaknesses in logical development or
presentation.
The researcher should also “see whether or not the
material, as it is presented, has unity and cohesion;
does the report stand upright and firm and exhibit a
definite pattern, like a marble arch? Or does it
resemble an old wall of moldering cement and loose
bricks.”
In addition the researcher should give due attention to
the fact that in his rough draft he has been consistent
or not. He/she should check the mechanics of writing
spelling and usage.
Preparation of the final bibliography. Next in order
comes the task of the preparation of the final
bibliography.
The bibliography, which is generally appended to the
research report, is a list of books in some way
pertinent to the research which has been done. It
should contain all those works which the researcher
has consulted.
The bibliography should be arranged alphabetically
and may be divided into two parts; the first part may
contain the names of books and pamphlets, and the
second part may contain the names of magazine and
newspaper articles.
Generally, this pattern of bibliography is considered
convenient and satisfactory from the point of view of
reader, though it is not the only way of presenting
bibliography.
Thank You