Professional Documents
Culture Documents
Primary Data Secondary Data
Primary Data Secondary Data
Secondary Data
Internal External
Published Secondary
Data
Computerized
Databases
Panels
Electronic scanner
Purchase Media services
Psychographic Advertising
General
& Lifestyles Evaluation
4-7
Audits
Descriptive Causal
Projective
Depth Interviews Techniques
Focus Groups
Word Association
In word association, respondents are presented with a list of
words, one at a time and asked to respond to each with the first
word that comes to mind. The words of interest, called test
words, are interspersed throughout the list which also contains
some neutral, or filler words to disguise the purpose of the
study. Responses are analyzed by calculating:
Completion Techniques
In Sentence completion, respondents are given incomplete
sentences and asked to complete them. Generally, they are
asked to use the first word or phrase that comes to mind.
Completion Techniques
In story completion, respondents are given part of
a story – enough to direct attention to a particular
topic but not to hint at the ending. They are
required to give the conclusion in their own words.
4-15
Construction Techniques
With a picture response, the respondents are
asked to describe a series of pictures of ordinary as
well as unusual events. The respondent's
interpretation of the pictures gives indications of that
individual's personality.
A Cartoon Test
Figure 5.4
Sears
Expressive Techniques
In expressive techniques, respondents are
presented with a verbal or visual situation and asked
to relate the feelings and attitudes of other people to
the situation.
Traditional Computer-Assisted
Mail Mail
Telephone Telephone
Interview Panel
Interviewing
Observation Methods 4-20
Classifying
Observation
Methods
Observation Methods
Concept of Causality
A statement such as "X causes Y " will have the
following meaning to an ordinary person and to a
scientist.
____________________________________________________
Ordinary Meaning Scientific Meaning
____________________________________________________
X is the only cause of Y. X is only one of a number of
possible causes of Y.
Experimental Design
An experimental design is a set of
procedures specifying
Validity in Experimentation
Internal validity refers to whether the
manipulation of the independent variables or
treatments actually caused the observed
effects on the dependent variables. Control
of extraneous variables is a necessary
condition for establishing internal validity.
External validity refers to whether the
cause-and-effect relationships found in the
experiment can be generalized. To what
populations, settings, times, independent
variables and dependent variables can the
results be projected?
4-28
Experimental Designs
Factorial Design
Is used to measure the effects of two or
more independent variables at various
levels.
A factorial design may also be
conceptualized as a table.
In a two-factor design, each level of
one variable represents a row and each
level of another variable represents a
column.
4-31
National Introduction
Interval Performance
Rating on a 8.2 9.1 9.6
0 to 10 Scale
Scaling Techniques
Comparative Noncomparative
Scales Scales
Semantic Stapel
Likert
Differential
4-37
Form
Brand Rank Order
1. Crest _________
2. Colgate _________
3. Aim _________
4. Gleem _________
5. Macleans _________
Form
Average Responses of Three Segments
Attribute Segment I Segment II Segment III
1. Mildness 8 2 4
2. Lather 2 4 17
3. Shrinkage 3 9 7
4. Price 53 17 9
5. Fragrance 9 0 19
6. Packaging 7 5 9
7. Moisturizing 5 3 20
8. Cleaning Power 13 60 15
Sum 100 100 100
4-40
Likert Scale
The Likert scale requires the respondents to indicate a degree of agreement or
disagreement with each of a series of statements about the stimulus objects.
SEARS IS:
Powerful --:--:--:--:-X-:--:--: Weak
Unreliable --:--:--:--:--:-X-:--: Reliable
Modern --:--:--:--:--:--:-X-: Old-fashioned
Stapel Scale
The Stapel scale is a unipolar rating scale with ten categories
numbered from -5 to +5, without a neutral point (zero). This scale
is usually presented vertically.
SEARS
+5 +5
+4 +4
+3 +3
+2 +2X
+1 +1
HIGH QUALITY POOR SERVICE
-1 -1
-2 -2
-3 -3
-4X -4
-5 -5
1 2 3 4 5
4-46
Validity
Construct validity addresses the question of what
construct or characteristic the scale is, in fact,
measuring. Construct validity includes convergent,
discriminant, and nomological validity.
Convergent validity is the extent to which the
scale correlates positively with other measures of the
same construct.
Discriminant validity is the extent to which a
measure does not correlate with other constructs
from which it is supposed to differ.
Nomological validity is the extent to which the
scale correlates in theoretically predicted ways with
measures of different but related constructs.
4-47
Questionnaire Definition
A questionnaire is a formalized set of questions for
obtaining information from respondents.
4-48
Unstructured Questions
Structured Questions
Multiple-Choice Questions
Dichotomous Questions
Yes No
Credit Cash
Yes
Other
No
Store Bank Other
Charge Charge Charge
Card Card Card
Pretesting
Pretesting refers to the testing of the questionnaire
on a small sample of respondents to identify and
eliminate potential problems.
Observational Forms
Department Store Project
Who: Purchasers, browsers, males, females, parents
with children, or children alone.
What: Products/brands considered, products/brands
purchased, size, price of package inspected, or
influence of children or other family members.
When: Day, hour, date of observation.
Research Studies
Table 11.2
Sampling Techniques
Nonprobability Probability
Sampling Techniques Sampling Techniques
Check Questionnaire
Edit
Code
Transcribe
Clean Data
Dependence Interdependence
Technique Technique
Frequency Distribution
In a frequency distribution, one variable is
considered at a time.
A frequency distribution for a variable produces a
table of frequency counts, percentages, and
cumulative percentages for all the values associated
with that variable.
Statistics Associated with Frequency Distribution 4-69
Measures of Location
The mean, or average value, is the most commonly used
measure of central tendency. The mean, X ,is given by
n
X = Σ X i /n
i=1
Where,
Xi = Observed values of the variable X
n = Number of observations (sample size)
Measures of Location
Measures of Variability
The range measures the spread of the data. It is
simply the difference between the largest and
smallest values in the sample. Range = Xlargest –
Xsmallest.
The interquartile range is the difference between
the 75th and 25th percentile. For a set of data
points arranged in order of magnitude, the pth
percentile is the value that has p% of the data points
below it and (100 - p)% above it.
Statistics Associated with Frequency Distribution 4-72
Measures of Variability
CV = s x/X
Statistics Associated with Frequency Distribution 4-73
Measures of Shape
Skewness. The tendency of the deviations from the
mean to be larger in one direction than in the other.
It can be thought of as the tendency for one tail of
the distribution to be heavier than the other.
Skewness of a Distribution
Figure 15.2
Symmetric Distribution
Skewed Distribution
Mean
Median
Mode
(a)
Tests of Tests of
Association Differences
Median/
Distributions Means Proportions
Rankings
4-77
Cross-Tabulation
While a frequency distribution describes one variable
at a time, a cross-tabulation describes two or more
variables simultaneously.
Cross-tabulation results in tables that reflect the joint
distribution of two or more variables with a limited
number of categories or distinct values, e.g., Table
15.3.
4-78
Gender
Row
Internet Usage Male Female Total
Light (1) 5 10 15
Heavy (2) 10 5 15
Column Total 15 15
4-79
Gender
Internet Usage
Tabulation
Fig. 15.7
Original Two Variables
No 35% 35%
Income
Eat Frequently in Fast- Low High
Food Restaurants
Family size Family size
Small Large Small Large
Yes 65% 65% 65% 65%
No 35% 35% 35% 35%
Column totals 100% 100% 100% 100%
Number of respondents 250 250 250 250
4-86
Chi-square Distribution
Figure 15.8
Do Not Reject
H0
Reject H0
χ2
Critical
Value
Statistics Associated with Cross-Tabulation 4-87
Chi-Square
Chi-Square
χ2 = Σ (f o - f e) 2
fe
all
cells
Statistics Associated with Cross-Tabulation 4-89
Chi-Square
χ 2
For the data in Table 15.3, the value of is
calculated as:
= 3.333
Statistics Associated with Cross-Tabulation 4-90
Lambda Coefficient
Asymmetric lambda measures the percentage
improvement in predicting the value of the dependent
variable, given the value of the independent variable.
Lambda also varies between 0 and 1. A value of 0 means
no improvement in prediction. A value of 1 indicates that
the prediction can be made without error. This happens
when each independent variable category is associated
with a single category of the dependent variable.
Asymmetric lambda is computed for each of the variables
(treating it as the dependent variable).
A symmetric lambda is also computed, which is a kind
of average of the two asymmetric values. The symmetric
lambda does not make an assumption about which
variable is dependent. It measures the overall
improvement when prediction is done in both directions.
A Classification of Hypothesis Testing
4-91
Non-Parametric Tests
Nonparametric tests are used when the independent
variables are nonmetric. Like parametric tests,
nonparametric tests are available for testing variables
from one sample, two independent samples, or two
related samples.
Non-Parametric Tests 4-93
One Sample
Sometimes the researcher wants to test whether the
observations for a particular variable could reasonably
have come from a particular distribution, such as the
normal, uniform, or Poisson distribution.
One Sample
The chi-square test can also be performed on a
single variable from one sample. In this context, the
chi-square serves as a goodness-of-fit test.
The runs test is a test of randomness for the
dichotomous variables. This test is conducted by
determining whether the order or sequence in which
observations are obtained is random.
The binomial test is also a goodness-of-fit test for
dichotomous variables. It tests the goodness of fit of
the observed number of observations in each
category to the number expected under a specified
binomial distribution.
Non-Parametric Tests 4-95
SPSS Windows
The main program in SPSS is FREQUENCIES. It
produces a table of frequency counts, percentages,
and cumulative percentages for the values of each
variable. It gives all of the associated statistics.
If the data are interval scaled and only the summary
statistics are desired, the DESCRIPTIVES procedure
can be used.
The EXPLORE procedure produces summary statistics
and graphical displays, either for all of the cases or
separately for groups of cases. Mean, median,
variance, standard deviation, minimum, maximum,
and range are some of the statistics that can be
calculated.
4-97
SPSS Windows
To select these procedures click:
Analyze>Descriptive Statistics>Frequencies
Analyze>Descriptive Statistics>Descriptives
Analyze>Descriptive Statistics>Explore
Analyze>Descriptive Statistics>Crosstabs
4-98
SPSS Windows
The major program for conducting parametric
tests in SPSS is COMPARE MEANS. This program can
be used to conduct t tests on one sample or
independent or paired samples. To select these
procedures using SPSS for Windows click:
Analyze>Compare Means>Means …
Analyze>Compare Means>One-Sample T Test …
Analyze>Compare Means>Independent-
Samples T Test …
Analyze>Compare Means>Paired-Samples T
Test …
4-99
SPSS Windows
The nonparametric tests discussed in this chapter can
be conducted using NONPARAMETRIC TESTS.
Analyze>Nonparametric Tests>Chi-Square …
Analyze>Nonparametric Tests>Binomial …
Analyze>Nonparametric Tests>Runs …
Analyze>Nonparametric Tests>1-Sample K-S …
Analyze>Nonparametric Tests>2 Independent
Samples …
Analyze>Nonparametric Tests>2 Related
Samples …
4-100
Regression Analysis
Regression coefficient. The estimated
parameter b is usually referred to as the non-
standardized regression coefficient.
Regression Analysis
Standardized regression coefficient. Also
termed the beta coefficient or beta weight, this is
the slope obtained by the regression of Y on X
when the data are standardized.
Multiple Regression
The general form of the multiple regression model
is as follows:
Y = β 0 + β 1 X1 + β 2 X2 + β 3 X3+ . . . + β k Xk + e
which is estimated by the following equation:
Multicollinearity
Multicollinearity arises when intercorrelations
among the predictors are very high.
Multicollinearity can result in several problems,
including:
The partial regression coefficients may not be
estimated precisely. The standard errors are likely
to be high.
The magnitudes as well as the signs of the partial
regression coefficients may change from sample
to sample.
It becomes difficult to assess the relative
importance of the independent variables in
explaining the variation in the dependent variable.
Predictor variables may be incorrectly included or
removed in stepwise regression.
4-108
SPSS Windows
The CORRELATE program computes Pearson product moment correlations
and partial correlations with significance levels. Univariate statistics,
covariance, and cross-product deviations may also be requested.
Significance levels are included in the output. To select these procedures
using SPSS for Windows click:
Analyze>Correlate>Bivariate …
Analyze>Correlate>Partial …
Graphs>Scatter …>Simple>Define
Analyze>Regression Linear …
Similarities and Differences between ANOVA,
4-109
Similarities
Number of One One One
dependent
variables
Number of
independent Multiple Multiple Multiple
variables
Differences
Nature of the
dependent Metric Metric Categorical
variables
Nature of the
independent Categorical Metric Metric
variables
4-110
Discriminant Analysis
Discriminant analysis is a technique for analyzing data
when the criterion or dependent variable is categorical
and the predictor or independent variables are interval in
nature.
SPSS Windows
The DISCRIMINANT program performs both two-
group and multiple discriminant analysis. To select
this procedure using SPSS for Windows click:
Analyze>Classify>Discriminant …
4-115
Factor Analysis
Factor analysis is a general name denoting a class of
procedures primarily used for data reduction and
summarization.
Factor analysis is an interdependence technique in that an
entire set of interdependent relationships is examined without
making the distinction between dependent and independent
variables.
Factor analysis is used in the following circumstances:
To identify underlying dimensions, or factors, that explain
Rotation of Factors
Interpretation of Factors
Calculation of Selection of
Factor Scores Surrogate Variables
SPSS Windows
Analyze>Data Reduction>Factor …
4-120
Cluster Analysis
Cluster analysis is a class of techniques used to
classify objects or cases into relatively homogeneous
groups called clusters. Objects in each cluster tend
to be similar to each other and dissimilar to objects in
the other clusters. Cluster analysis is also called
classification analysis, or numerical taxonomy.
Both cluster analysis and discriminant analysis are
concerned with classification. However, discriminant
analysis requires prior knowledge of the cluster or
group membership for each object or case included,
to develop the classification rule. In contrast, in
cluster analysis there is no a priori information about
the group or cluster membership for any of the
objects. Groups or clusters are suggested by the
data, not defined a priori.
4-121
Variable 1
Variable 2
4-122
Hierarchical Nonhierarchical
Agglomerative Divisive
Ward’s Method
Cluster 1 Cluster 2
Complete Linkage
Maximum
Distance
Cluster 1 Cluster 2
Average Linkage
Average Distance
Cluster 1 Cluster 2
4-127
Centroid Method
4-128
SPSS Windows
To select this procedures using SPSS for Windows click:
Analyze>Classify>Hierarchical Cluster …
Analyze>Classify>K-Means Cluster …