Professional Documents
Culture Documents
Machine Learning For Business Analytics: Concepts, Techniques and Applications With JMP Pro, 2nd Edition Galit Shmueli
Machine Learning For Business Analytics: Concepts, Techniques and Applications With JMP Pro, 2nd Edition Galit Shmueli
Machine Learning For Business Analytics: Concepts, Techniques and Applications With JMP Pro, 2nd Edition Galit Shmueli
https://ebookmeta.com/product/machine-learning-for-business-
analytics-concepts-techniques-and-applications-with-jmp-pro-2nd-
edition-galit-shmueli-2/
https://ebookmeta.com/product/machine-learning-for-intelligent-
multimedia-analytics-techniques-and-applications-pardeep-kumar-
amit-kumar-singh-eds/
https://ebookmeta.com/product/machine-learning-techniques-and-
analytics-for-cloud-security-advances-in-learning-analytics-for-
intelligent-cloud-iot-systems-1st-edition-chakraborty/
Machine Learning for Intelligent Multimedia Analytics
Techniques and Applications Studies in Big Data 82
Pardeep Kumar Editor Amit Kumar Singh Editor
https://ebookmeta.com/product/machine-learning-for-intelligent-
multimedia-analytics-techniques-and-applications-studies-in-big-
data-82-pardeep-kumar-editor-amit-kumar-singh-editor/
https://ebookmeta.com/product/graph-powered-analytics-and-
machine-learning-with-tigergraph-driving-business-outcomes-with-
connected-data-1st-edition-victor-lee/
https://ebookmeta.com/product/fundamentals-of-predictive-
analytics-with-jmp-3rd-edition-klimberg/
https://ebookmeta.com/product/cambridge-igcse-and-o-level-
history-workbook-2c-depth-study-the-united-states-1919-41-2nd-
edition-benjamin-harrison/
https://ebookmeta.com/product/data-science-analytics-and-machine-
learning-with-r-1st-edition-luiz-favero/
Table of Contents
COVER
TITLE PAGE
COPYRIGHT
FOREWORD
PREFACE
NOTES
ACKNOWLEDGMENTS
PART I: PRELIMINARIES
1 INTRODUCTION
1.1 WHAT IS BUSINESS ANALYTICS?
1.2 WHAT IS MACHINE LEARNING?
1.3 MACHINE LEARNING, AI, AND RELATED TERMS
1.4 BIG DATA
1.5 DATA SCIENCE
1.6 WHY ARE THERE SO MANY DIFFERENT METHODS?
1.7 TERMINOLOGY AND NOTATION
1.8 ROAD MAPS TO THIS BOOK
2 OVERVIEW OF THE MACHINE LEARNING PROCESS
2.1 INTRODUCTION
2.2 CORE IDEAS IN MACHINE LEARNING
2.3 THE STEPS IN A MACHINE LEARNING PROJECT
2.4 PRELIMINARY STEPS
2.5 PREDICTIVE POWER AND OVERFITTING
2.6 BUILDING A PREDICTIVE MODEL WITH JMP Pro
2.7 USING JMP Pro FOR MACHINE LEARNING
2.8 AUTOMATING MACHINE LEARNING SOLUTIONS
2.9 ETHICAL PRACTICE IN MACHINE LEARNING
NOTES
PART II: DATA EXPLORATION AND DIMENSION REDUCTION
3 DATA VISUALIZATION
3.1 INTRODUCTION
3.2 DATA EXAMPLES
3.3 BASIC CHARTS: BAR CHARTS, LINE GRAPHS, AND SCATTER
PLOTS
3.4 MULTIDIMENSIONAL VISUALIZATION
3.5 SPECIALIZED VISUALIZATIONS
3.6 SUMMARY: MAJOR VISUALIZATIONS AND OPERATIONS,
ACCORDING TO MACHINE LEARNING GOAL
NOTES
4 DIMENSION REDUCTION
4.1 INTRODUCTION
4.2 CURSE OF DIMENSIONALITY
4.3 PRACTICAL CONSIDERATIONS
4.4 DATA SUMMARIES
4.5 CORRELATION ANALYSIS
4.6 REDUCING THE NUMBER OF CATEGORIES IN CATEGORICAL
VARIABLES
4.7 CONVERTING A CATEGORICAL VARIABLE TO A CONTINUOUS
VARIABLE
4.8 PRINCIPAL COMPONENT ANALYSIS
4.9 DIMENSION REDUCTION USING REGRESSION MODELS
4.10 DIMENSION REDUCTION USING CLASSIFICATION AND
REGRESSION TREES
NOTES
PART III: PERFORMANCE EVALUATION
5 EVALUATING PREDICTIVE PERFORMANCE
5.1 INTRODUCTION
5.2 EVALUATING PREDICTIVE PERFORMANCE
5.3 JUDGING CLASSIFIER PERFORMANCE
5.4 JUDGING RANKING PERFORMANCE
5.5 OVERSAMPLING
PART IV: PREDICTION AND CLASSIFICATION METHODS
6 MULTIPLE LINEAR REGRESSION
6.1 INTRODUCTION
6.2 EXPLANATORY VS. PREDICTIVE MODELING
6.3 ESTIMATING THE REGRESSION EQUATION AND PREDICTION
6.4 VARIABLE SELECTION IN LINEAR REGRESSION
NOTES
7 k‐NEAREST NEIGHBORS (k‐NN)
7.1 THE k‐NN CLASSIFIER (CATEGORICAL OUTCOME)
7.2 K‐NN FOR A NUMERICAL RESPONSE
7.3 ADVANTAGES AND SHORTCOMINGS OF K‐NN ALGORITHMS
NOTES
8 THE NAIVE BAYES CLASSIFIER
8.1 INTRODUCTION
8.2 APPLYING THE FULL (EXACT) BAYESIAN CLASSIFIER
8.3 SOLUTION: NAIVE BAYES
8.4 ADVANTAGES AND SHORTCOMINGS OF THE NAIVE BAYES
CLASSIFIER
9 CLASSIFICATION AND REGRESSION TREES
9.1 INTRODUCTION
9.2 CLASSIFICATION TREES
9.3 GROWING A TREE FOR RIDING MOWERS EXAMPLE
9.4 EVALUATING THE PERFORMANCE OF A CLASSIFICATION TREE
9.5 AVOIDING OVERFITTING
9.6 CLASSIFICATION RULES FROM TREES
9.7 CLASSIFICATION TREES FOR MORE THAN TWO CLASSES
9.8 REGRESSION TREES
9.9 ADVANTAGES AND WEAKNESSES OF A SINGLE TREE
9.10 IMPROVING PREDICTION: RANDOM FORESTS AND BOOSTED
TREES
NOTES
10 LOGISTIC REGRESSION
10.1 INTRODUCTION
10.2 THE LOGISTIC REGRESSION MODEL
10.3 EXAMPLE: ACCEPTANCE OF PERSONAL LOAN
10.4 EVALUATING CLASSIFICATION PERFORMANCE
10.5 VARIABLE SELECTION
10.6 LOGISTIC REGRESSION FOR MULTI‐CLASS CLASSIFICATION
10.7 EXAMPLE OF COMPLETE ANALYSIS: PREDICTING DELAYED
FLIGHTS
Notes
11 NEURAL NETS
11.1 INTRODUCTION
11.2 CONCEPT AND STRUCTURE OF A NEURAL NETWORK
11.3 FITTING A NETWORK TO DATA
11.4 USER INPUT IN JMP Pro
11.5 EXPLORING THE RELATIONSHIP BETWEEN PREDICTORS AND
OUTCOME
11.6 DEEP LEARNING
11.7 ADVANTAGES AND WEAKNESSES OF NEURAL NETWORKS
NOTES
12 DISCRIMINANT ANALYSIS
12.1 INTRODUCTION
12.2 DISTANCE OF AN OBSERVATION FROM A CLASS
12.3 FROM DISTANCES TO PROPENSITIES AND CLASSIFICATIONS
12.4 CLASSIFICATION PERFORMANCE OF DISCRIMINANT ANALYSIS
12.5 PRIOR PROBABILITIES
12.6 CLASSIFYING MORE THAN TWO CLASSES
12.7 ADVANTAGES AND WEAKNESSES
NOTES
13 GENERATING, COMPARING, AND COMBINING MULTIPLE
MODELS
13.1 ENSEMBLES
13.2 AUTOMATED MACHINE LEARNING (AUTOML)
13.3 SUMMARY
NOTE
PART V: INTERVENTION AND USER FEEDBACK
14 INTERVENTIONS: EXPERIMENTS, UPLIFT MODELS, AND
REINFORCEMENT LEARNING
14.1 INTRODUCTION
14.2 A/B TESTING
14.3 UPLIFT (PERSUASION) MODELING
14.4 REINFORCEMENT LEARNING
14.5 SUMMARY
NOTES
PART VI: MINING RELATIONSHIPS AMONG RECORDS
15 ASSOCIATION RULES AND COLLABORATIVE FILTERING
15.1 ASSOCIATION RULES
15.2 COLLABORATIVE FILTERING
15.3 SUMMARY
NOTES
16 CLUSTER ANALYSIS
16.1 INTRODUCTION
16.2 MEASURING DISTANCE BETWEEN TWO RECORDS
16.3 MEASURING DISTANCE BETWEEN TWO CLUSTERS
16.4 HIERARCHICAL (AGGLOMERATIVE) CLUSTERING
16.5 NONHIERARCHICAL CLUSTERING: THE K‐MEANS ALGORITHM
NOTE
PART VII: FORECASTING TIME SERIES
17 HANDLING TIME SERIES
17.1 INTRODUCTION
17.2 DESCRIPTIVE VS. PREDICTIVE MODELING
17.3 POPULAR FORECASTING METHODS IN BUSINESS
17.4 TIME SERIES COMPONENTS
17.5 DATA PARTITIONING AND PERFORMANCE EVALUATION
NOTES
18 REGRESSION‐BASED FORECASTING
18.1 A MODEL WITH TREND
18.2 A MODEL WITH SEASONALITY
18.3 A MODEL WITH TREND AND SEASONALITY
18.4 AUTOCORRELATION AND ARIMA MODELS
Notes
19 SMOOTHING AND DEEP LEARNING METHODS FOR
FORECASTING
19.1 INTRODUCTION
19.2 MOVING AVERAGE
19.3 SIMPLE EXPONENTIAL SMOOTHING
19.4 ADVANCED EXPONENTIAL SMOOTHING
19.5 DEEP LEARNING FOR FORECASTING
NOTES
PART VIII: DATA ANALYTICS
20 TEXT MINING
20.1 INTRODUCTION
20.2 THE TABULAR REPRESENTATION OF TEXT: DOCUMENT–TERM
MATRIX AND “BAG‐OF‐WORDS”
20.3 BAG‐OF‐WORDS VS. MEANING EXTRACTION AT DOCUMENT
LEVEL
20.4 PREPROCESSING THE TEXT
20.5 IMPLEMENTING MACHINE LEARNING METHODS
20.6 EXAMPLE: ONLINE DISCUSSIONS ON AUTOS AND
ELECTRONICS
20.7 EXAMPLE: SENTIMENT ANALYSIS OF MOVIE REVIEWS
20.8 SUMMARY
NOTES
21 RESPONSIBLE DATA SCIENCE
21.1 INTRODUCTION
21.2 UNINTENTIONAL HARM
21.3 LEGAL CONSIDERATIONS
21.4 PRINCIPLES OF RESPONSIBLE DATA SCIENCE
21.5 A RESPONSIBLE DATA SCIENCE FRAMEWORK
21.6 DOCUMENTATION TOOLS
21.7 EXAMPLE: APPLYING THE RDS FRAMEWORK TO THE COMPAS
EXAMPLE
21.8 SUMMARY
NOTES
PART IX: CASES
22 CASES
22.1 CHARLES BOOK CLUB
22.2 GERMAN CREDIT
22.3 TAYKO SOFTWARE CATALOGER
22.4 POLITICAL PERSUASION
22.5 TAXI CANCELLATIONS
22.6 SEGMENTING CONSUMERS OF BATH SOAP
22.7 CATALOG CROSS‐SELLING
22.8 DIRECT‐MAIL FUNDRAISING
22.9 TIME SERIES CASE: FORECASTING PUBLIC TRANSPORTATION
DEMAND
22.10 LOAN APPROVAL
NOTES
REFERENCES
DATA FILES USED IN THE BOOK
INDEX
END USER LICENSE AGREEMENT
List of Tables
Chapter 1
TABLE 1.1 Organization of machine learning methods in this book,
According t...
Chapter 2
TABLE 2.1 Hypothetical Data on Advertising Expenditures and
Subsequent Sales
TABLE 2.2 First 10 Records in the West Roxbury home values dataset
TABLE 2.3 Description of Variables in West Roxbury (Boston) Home
Value Datas...
TABLE 2.4 Outlier in West Roxbury data
TABLE 2.5 Sample from a database of credit applications
TABLE 2.6 Sample from a bank database
TABLE 2.7
Chapter 3
TABLE 3.1 Description of Variables in the Boston Housing Dataset
Chapter 4
TABLE 4.1 Description of Variables in the Boston Housing Dataset
TABLE 4.2 Sample from the 77 breakfast cereals dataset
TABLE 4.3 Description of the Variables in the Breakfast Cereal
Dataset
TABLE 4.4 Cereal Calories and Ratings
Chapter 5
TABLE 5.1 Classification Matrix: Meaning of Each Cell
Chapter 6
TABLE 6.1 Variables in the Toyota Corolla Example
TABLE 6.2 Description of Variables for Boston Housing Example
TABLE 6.3 Description of Variables for Tayko Software Example
TABLE 6.4 Description of Variables for Airfare Example
Chapter 7
TABLE 7.1 Lot size, income, and ownership of a riding mower for 24
Household...
Chapter 8
TABLE 8.1 Tabular Summary of Financial Reporting Data
TABLE 8.2 Information on 10 Companies
TABLE 8.3 Description of Variables for Flight Delays Example
Chapter 9
TABLE 9.1 Lot Size, Income, and Ownership of a Riding Mower for
24 Household...
Chapter 10
TABLE 10.1 Description of Predictors for Acceptance of Personal
Loan Example...
TABLE 10.2 Description of Predictors for Flight Delays Example
Chapter 11
TABLE 11.1 Tiny Example on Tasting Scores for Six Consumers and
Two Predicto...
TABLE 11.2 Description of Variables for Automobile Accident
Example
TABLE 11.3 Data for Credit Card Example and Variable Descriptions
Chapter 14
TABLE 14.1 Example of raw data resulting from A/B test
TABLE 14.2 Results from A/B test at PhoToTo, evaluating the effect
of a new ...
TABLE 14.3 Data on voters (small subset of variables and records)
and Data D...
TABLE 14.4 Results of sending a pro‐Democratic message to voters
TABLE 14.5 Target variable (Moved_AD) and treatment variable
(Message) added...
TABLE 14.6 Classifications and propensities from predictive model
(small ext...
TABLE 14.7 Classifications and propensities from predictive model
(small ext...
TABLE 14.8 Uplift: change in propensities from sending message vs.
not sendi...
Chapter 15
TABLE 15.1 Transactions Database for Purchases of Different‐Colored
Cellular...
TABLE 15.2 Phone Faceplate Data in Binary Incidence Matrix Format
TABLE 15.3 Example for Computing Conviction
TABLE 15.4 Fifty Transactions of Randomly Assigned Items
TABLE 15.5 Schematic of matrix format with ratings data
TABLE 15.6 Sample of records from the Netflix Prize contest, for a
subset of...
TABLE 15.7 Sample of data on satellite radio customers
TABLE 15.8 Data on purchases of online statistics courses
TABLE 15.9 Excerpt from data on cosmetics purchases in binary
matrix form
Chapter 16
TABLE 16.1 Data on 22 Public Utilities
TABLE 16.2 Original and Standardized Measurements for Sales and
Fuel Cost
TABLE 16.3 Distance Matrix Between Pairs of the First Five Utilities,
Using ...
TABLE 16.4 Distance Matrix After Arizona and Commonwealth
Consolidation Clus...
TABLE 16.5 Distance of Each record from Each Centroid
TABLE 16.6 Distance of Each record from Each Newly Calculated
Centroid
Chapter 18
TABLE 18.1 Output variable (Ridership) and predictor variable (t)
used to fi...
TABLE 18.2 New categorical variable (Season) to be used in a linear
regressi...
TABLE 18.3 First 24 months of Amtrak ridership series
Chapter 19
TABLE 19.1 Converting a time series of length t into a sequence of
sub‐serie...
Chapter 21
TABLE 21.1 Logistic regression model for COMPAS data
TABLE 21.2 Jigsaw toxicity scores for certain phrases
Chapter 22
TABLE 22.1 List of Variables in Charles Book Club Dataset
TABLE 22.2 Average Net Profit (Deutsche Marks)
TABLE 22.3 Variables for the German Credit Dataset
TABLE 22.4 Description of Variables for Tayko Dataset
TABLE 22.5 Description of variables for each household
TABLE 22.6 Description of Variables for the Fundraising Dataset
List of Illustrations
Chapter 1
FIGURE 1.1 Two methods for separating owners from non‐owners
FIGURE 1.2 Machine learning from a process perspective. Numbers
in parenthes...
FIGURE 1.3 JMP home window (top) and data table (bottom) (on a
Windows PC)...
FIGURE 1.4 JMP starter window (on a Windows PC)
Chapter 2
FIGURE 2.1 Schematic of the machine learning process
FIGURE 2.2 Four JMP modeling types
FIGURE 2.3 Example dataset, showing the columns panel (on the
left) and the ...
FIGURE 2.4 Scatterplot for expenditure and revenue data
FIGURE 2.5 Overfitting: This function fits the data with no errors
FIGURE 2.6 Three data partitions and their role in the machine
learning proc...
FIGURE 2.7 Partitioning the data using the JMP Pro Make Validation
Column util...
FIGURE 2.8 Using the JMP Pro Fit Model platform for multiple linear
regression
FIGURE 2.9 Use the red triangle in the fit least squares analysis
window to sa...
FIGURE 2.10 Predictions for the first twenty records in the West
Roxbury housi...
FIGURE 2.11 Summary statistics for error rates for training (top) and
validati...
FIGURE 2.12 Crossvalidation statistics from the JMP Pro analysis
window for th...
FIGURE 2.13 Two records scored using the saved linear regression
model
FIGURE 2.14 Layers of tools supporting machine learning
automation
Chapter 3
FIGURE 3.1 First 10 records in the Boston housing dataset
FIGURE 3.2 The JMP Graph Builder initial window
FIGURE 3.3 Basic plots: line chart (top left), scatter plot (top right),
bar...
FIGURE 3.4 Distribution charts for numerical variable MEDV:
histogram and bo...
FIGURE 3.5 Side‐by‐side boxplots for exploring the CAT.MEDV
outcome variable...
FIGURE 3.6 Color‐coded correlation table and heatmap (also called a
color ma...
FIGURE 3.7 Missing data pattern table. The pattern indicates which
columns a...
FIGURE 3.8 Missing data pattern cell plot
FIGURE 3.9 Color‐coding and trellising by a categorical variable.
Left: scat...
FIGURE 3.10 Scatter plot of NOX vs. LSTAT, trellised by RAD
FIGURE 3.11 Scatter plot matrix for MEDV and three numerical
predictors. The...
FIGURE 3.12 Rescaling can enhance plots and reveal patterns:
original scale ...
FIGURE 3.13 Time series line charts using different aggregations
(right pane...
FIGURE 3.14 Using the Local Data Filter in the Graph Builder to view
only da...
FIGURE 3.15 Adding a trendline in Graph Builder
FIGURE 3.16 Scatter plot with labeled points
FIGURE 3.17 Scatter plot of 5000 records with reduced marker size,
different...
FIGURE 3.18 Parallel coordinates plot for Boston Housing data. Each
line is ...
FIGURE 3.19 Multiple inter‐linked plots in a single dashboard (a JMP
dashboard...
FIGURE 3.20 Network graph of eBay sellers (left side) and buyers
(right side...
FIGURE 3.21 Treemap showing nearly 11,000 eBay auctions,
organized by item c...
FIGURE 3.22 Aircraft incidents in the USA. Each dot is an incident,
colored ...
FIGURE 3.23 World maps comparing wellbeing (Top) to GDP
(bottom). Shading by...
Chapter 4
FIGURE 4.1 First 10 records in the Boston Housing data
FIGURE 4.2 Summary statistics for the Boston Housing data from the
Columns V...
FIGURE 4.3 Basic graphs and summary statistics for CRIM and CHAS
from the Di...
FIGURE 4.4 Correlation table for Boston Housing data, generated
using the Mu...
FIGURE 4.5 The initial Tabulate window. Drag and drop variables in
the drop ...
FIGURE 4.6 Tabulations for the Boston Housing data, created using
Tabulate ...
FIGURE 4.7 Distribution of CAT.MEDV (red indicates CAT.MEDV = 1)
by ZN. Simi...
FIGURE 4.8 Quarterly revenues of Toys “R” Us, 1992–1995...
FIGURE 4.9 Combining categories using Recode
FIGURE 4.10 Correlation analysis, scatterplot matrix, and descriptive
statis...
FIGURE 4.11 Scatterplot of rating vs. calories for 77 breakfast
cereals, wit...
FIGURE 4.12 Initial launch dialog and output from principal
component analys...
FIGURE 4.13 Principal component scores from PCA of calories and
rating for t...
FIGURE 4.14 PCA output using all 13 numerical variables in the
breakfast cer...
FIGURE 4.15 PCA output using all 13 numerical variables
(standardized) in th...
FIGURE 4.16 Score scatter plot of the second vs. first principal
components ...
FIGURE 4.17 Principal components of standardized (left) and
unscaled (right) w...
Chapter 5
FIGURE 5.1 Prediction error metrics from a model for Toyota car
prices. Trai...
FIGURE 5.2 Histograms, boxplots, and summary statistics of Toyota
price pred...
FIGURE 5.3 High (top) and low (bottom) levels of separation
between two clas...
FIGURE 5.4 Classification matrix based on 3000 observations and
two classes...
FIGURE 5.5 Riding mower example: 24 records with the actual class
and probab...
FIGURE 5.6 Classification matrices based on thresholds of 0.5, 0.25,
and 0.7...
FIGURE 5.7 Decision Threshold report, with fitted probability graph,
confusi...
FIGURE 5.8 Using the Formula Editor with Conditional and
Comparison function...
FIGURE 5.9 ROC curve and AUC (Area) for the riding mower
example
FIGURE 5.10 Specifying misclassification costs in JMP using the Profit
Matri...
FIGURE 5.11 Illustration of lift calculations for the riding mower
example (...
FIGURE 5.12 Lift curve for the riding mower example
FIGURE 5.13 Hypothetical example, lift curve incorporating costs
FIGURE 5.14 Classification assuming equal costs of misclassification
FIGURE 5.15 Classification assuming unequal costs of
misclassification
FIGURE 5.16 Classification using oversampling to account for
unequal costs...
FIGURE 5.17 Formula to adjust for oversampling
FIGURE 5.18 Lift curve for transaction data
Chapter 6
FIGURE 6.1 Prices and attributes for sample of 15 cars
FIGURE 6.2 Estimated coefficients for regression model of price vs.
car vari...
FIGURE 6.3 Performance of regression model on training and
validation sets...
FIGURE 6.4 Predicted prices and residuals (errors) for the first 15
cars in ...
FIGURE 6.5 Prediction formula for price, saved to the data table
FIGURE 6.6 Histogram, boxplot, and summary statistics for model
residuals (b...
FIGURE 6.7 Prediction profiler for price model
FIGURE 6.8 Effect summary table for price
FIGURE 6.9 Selection of the Stepwise personality from the Fit Model
dialog w...
FIGURE 6.10 Exhaustive search (all possible models) result for
reducing pred...
FIGURE 6.11 Forward selection results in Toyota Corolla price
example, using...
FIGURE 6.12 Backward elimination results using the Max Validation
RSquare st...
FIGURE 6.13 Backward elimination results using the P‐value
Threshold stoppin...
FIGURE 6.14 Generalized Regression option (personality) in the Fit
Model spe...
FIGURE 6.15 Lasso regression applied to the Toyota Corolla data
FIGURE 6.16 Ridge regression applied to the Toyota Corolla data
FIGURE 6.17 Elastic net regression applied to the Toyota Corolla
data
FIGURE 6.18 Comparison of lasso, ridge, elastic net, and least
squares regre...
Chapter 7
FIGURE 7.1 Scatterplot of Lot Size vs. Income for the 18 households
in the t...
FIGURE 7.2 Selection of the k‐Nearest Neighbors option from the
Analyze > Pr...
FIGURE 7.3 Model selection graph, holdout misclassification rates for
variou...
FIGURE 7.4 Classifying a new household using the “best k”= 8
Chapter 8
FIGURE 8.1 The Naive Bayes launch dialog
FIGURE 8.2 Tabulation of airport training data showing conditional
probabili...
FIGURE 8.3 Confusion matrix, ROC curves, and lift curves
FIGURE 8.4 Descriptive statistics for Distance as a continuous
predictor for...
Chapter 9
FIGURE 9.1 Example of a tree for classifying bank customers as loan
acceptor...
FIGURE 9.2 Scatterplot of Lot Size vs. Income for 24 owners and
non‐owners o...
FIGURE 9.3 Splitting the 24 records by income value of 85.5
FIGURE 9.4 The initial recursive partition output in JMP
FIGURE 9.5 JMP recursive partition output after 1 split. Display
Options Sh...
FIGURE 9.6 JMP recursive partition output for full tree
FIGURE 9.7 JMP leaf report for full tree
FIGURE 9.8 Riding mower data table with saved prediction formula
and classif...
FIGURE 9.9 Sample of 20 customers of universal bank
FIGURE 9.10 A full tree for the universal bank data using the training
set (...
FIGURE 9.11 Tree view of first 10 splits for the universal bank data
using t...
FIGURE 9.12 Universal bank data: classification matrix and error
rates for t...
FIGURE 9.13 Error rate as a function of the number of splits for
training vs...
FIGURE 9.14 Fit details and split history for the training (blue line),
vali...
FIGURE 9.15 Universal bank data: leaf report for pruned tree
FIGURE 9.16 Regression tree for Toyota Corolla prices after one split
FIGURE 9.17 Final regression tree for Toyota Corolla prices, small
tree view
FIGURE 9.18 Predictor (column) contributions and leaf report for
final regre...
FIGURE 9.19 Scatterplot describing a two‐predictor case with two
classes. Th...
FIGURE 9.20 Universal bank data: confusion matrix and error rates
for the tr...
FIGURE 9.21 Options for Bootstrap Forest and Boosted Tree in JMP
Pro
Chapter 10
FIGURE 10.1 Odds (top) and logit (bottom) as a function of p
FIGURE 10.2 The logistic curve for personal loan as a function of
income. A ...
FIGURE 10.3 Logistic regression parameter estimates (coefficients)
table for...
FIGURE 10.4 Estimated logistic equation for the Universal Bank data
based on...
FIGURE 10.5 The equation for calculating the probability of
acceptance, Prob...
FIGURE 10.6 Scoring the data: Output for the first 15 Universal Bank
custome...
FIGURE 10.7 Decision threshold report for Universal Bank loan offer
FIGURE 10.8 Lift curve of validation data for Universal Bank loan
offer
FIGURE 10.9 Parameter estimates for ordinal logistic regression
model for ac...
FIGURE 10.10 Parameter estimates for nominal logistic regression
model for a...
FIGURE 10.11 Sample of 20 flights
FIGURE 10.12 Proportion of delayed flights by each of the
predictors. The mo...
FIGURE 10.13 Heatmap of proportion of delayed flights (darker =
higher propo...
FIGURE 10.14 Estimated logistic regression model for delayed flights
with on...
FIGURE 10.15 Logit for delayed flights with one predictor,
Destination
FIGURE 10.16 Estimated logistic regression model for delayed flights
with al...
FIGURE 10.17 Prediction profiler for flight delays. The contours show
how th...
FIGURE 10.18 Fit statistics, including misclassification rates, for the
flig...
FIGURE 10.19 Lift curves and confusion matrices for the flight delays
data...
FIGURE 10.20 Effect Summary table, after the removal of Origin
FIGURE 10.21 Using Recode, from the Cols Utilities menu to group
day of the...
FIGURE 10.22 Output for logistic regression model with only seven
predictors...
FIGURE 10.23 Prediction Profiler showing best (bottom) and worst
(top) condi...
Chapter 11
FIGURE 11.1 Multilayer feedforward neural network
FIGURE 11.2 Neural network for the tiny example. Circles represent
nodes (“n...
FIGURE 11.3 Computing node outputs (values are to the right of
each node), u...
FIGURE 11.4 Output for neural network applied to the tiny data
example. Esti...
FIGURE 11.5 Saved formulas and classifications for the tiny data
example
FIGURE 11.6 Formula to compute output from the first hidden layer
node
FIGURE 11.7 Formula to compute Prob(Acceptance=like) in JMP
FIGURE 11.8 Subset from the accident data, for a high‐fatality
region...
FIGURE 11.9 Neural network models for accident data, with five
nodes in one ...
FIGURE 11.10 JMP Pro neural network for accident data with five
nodes and no...
FIGURE 11.11 JMP Pro Neural Model Launch dialog
FIGURE 11.12 Line drawing, from a 1893 Funk and Wagnalls
publication
FIGURE 11.13 Focusing on the line of the man's chin
FIGURE 11.14 pixel representation of line on man's chin using
shading (a) ...
FIGURE 11.15 Convolution network process, supervised learning: the
repeated ...
FIGURE 11.16 Autoencoder network process: the repeated filtering
in the netw...
Chapter 12
FIGURE 12.1 Scatterplot of Lot Size vs. Income for 24 owners and
non‐owners ...
FIGURE 12.2 Personal loan acceptance as a function of income and
credit card...
FIGURE 12.3 Discriminant analysis output for riding mower data,
displaying t...
FIGURE 12.4 Squared distances for riding mower data. To view this
table in J...
FIGURE 12.5 Output for riding mower data, displaying the
propensities (estim...
FIGURE 12.6 Class separation obtained from the discriminant model
(compared ...
FIGURE 12.7 Score summaries and confusion matrix for mower data
FIGURE 12.8 Discriminant scores for mower data, with the prior
probability f...
FIGURE 12.9 Sample of 15 automobile accidents from the 2001
Department of Tr...
FIGURE 12.10 Discriminant scores for the first 15 observations of the
accide...
FIGURE 12.11 Score Summaries (with confusion matrices),
probabilities, and s...
FIGURE 12.12 Formula for the squared distance for fatal accidents,
saved to ...
FIGURE 12.13 Formula for the probability of fatal accidents, saved to
the da...
Chapter 13
FIGURE 13.1 Validation ROC curves (zooming in to top left part) and
AUC valu...
FIGURE 13.2 Workflow Builder example, with recorded data
exploration and pre...
FIGURE 13.3 Model Screening platform JMP Pro (personal loan data)
FIGURE 13.4 Validation results for multiple models fit using Model
Screening...
Chapter 14
FIGURE 14.1 Test of hypothesis for two sample means (above) and
two sample p...
FIGURE 14.2 Uplift model specification dialog in JMP Pro (top) and
initial u...
FIGURE 14.3 Schematic of MDP, where the software agent “learns”
the human us...
Chapter 15
FIGURE 15.1 Recommendations under “frequently bought together”
are based on ...
FIGURE 15.2 Itemsets with support percentage of at least 20%
FIGURE 15.3 Data format and association analysis dialog window in
JMP Pro...
FIGURE 15.4 Frequent itemsets with minimum support 20% and
association rules...
FIGURE 15.5 Association rules output for random data
FIGURE 15.6 Subset of book purchase transactions in binary matrix
format (ri...
FIGURE 15.7 Association rules for book purchase transactions sorted
by lift...
FIGURE 15.8 Association rules for cosmetics purchases data
Chapter 16
FIGURE 16.1 Scatterplot of Fuel Cost vs. Sales for the 22 utilities
FIGURE 16.2 Two‐dimensional representation of several different
distance mea...
FIGURE 16.3 The hierarchical clustering dialog for the utilities data,
showi...
FIGURE 16.4 Dendrogram: single linkage for all 22 utilities, using all
8 mea...
FIGURE 16.5 Dendrogram: average linkage for all 22 utilities, using
all 8 me...
FIGURE 16.6 Cluster Summary for hierarchical clustering with six
clusters...
FIGURE 16.7 Parallel plot for the 22 utilities, from average linking
with 6 ...
FIGURE 16.8 Exploring the saved clusters using the Graph Builder
and the Col...
FIGURE 16.9 Color map for the 22 utilities. Rows are sorted by the 6
cluster...
FIGURE 16.10 k‐Means cluster dialog (top) and control panel
(bottom) for spe...
FIGURE 16.11 Output for k‐means clustering of 22 utilities into 6
clusters (...
FIGURE 16.12 Graph of standardized distances vs. cluster number
from the Gra...
FIGURE 16.13 Parallel coordinates (profile) plot of cluster centroids
for k‐...
FIGURE 16.14 Summary of cluster centroids, created using the
saved clusters ...
FIGURE 16.15 Biplot (top) and scatterplot matrix (bottom)
FIGURE 16.16 Initial output for k‐means clustering of 22 utilities,
showing ...
FIGURE 16.17 Comparing different choices of k in terms of the
standard devia...
Chapter 17
FIGURE 17.1 Monthly ridership on Amtrak trains (in thousands) from
January 1...
FIGURE 17.2 Time plots of the daily number of vehicles passing
through the B...
FIGURE 17.3 Plots that enhance the different components of the
time series. ...
FIGURE 17.4 Time Series analysis launch window
FIGURE 17.5 Naive and seasonal naive forecasts in a three‐year
validation se...
FIGURE 17.6 Average annual weekly hours spent by Canadian
manufacturing worker...
Chapter 18
FIGURE 18.1 A linear trend fit to Amtrak ridership
FIGURE 18.2 Fitted regression model with linear trend, regression
output, an...
FIGURE 18.3 Fitted trend and residual plot for Amtrak ridership data
produce...
FIGURE 18.4 Output from regression model with exponential trend,
fit to trai...
FIGURE 18.5 Adjusting forecasts of log(Ridership) to the original
scale and ...
FIGURE 18.6 Regression model with quadratic trend fit to the
training data...
FIGURE 18.7 Fitted regression model with seasonality. Parameter
estimates an...
FIGURE 18.8 Regression model with monthly (additive) seasonality
and quadrat...
FIGURE 18.9 JMP Time Series platform output for the first 24
months of the A...
FIGURE 18.10 Amtrak data: analysis of residuals from regression
model with s...
FIGURE 18.11 Partial output from of AR(1) model fitted to the
residual serie...
FIGURE 18.12 Forecasted residual
FIGURE 18.13 Autocorrelations of lag‐1 differenced S&P500 monthly
closing pr...
FIGURE 18.14 Partial output for AR(1) model on S&P500 monthly
closing prices...
FIGURE 18.15 Seasonally adjusted pre‐September‐11 AIR RPM
series
FIGURE 18.16 Average annual weekly hours spent by Canadian
manufacturing wor...
FIGURE 18.17 Quarterly revenues of Toys “R” Us, 1992–1995
FIGURE 18.18 Output for regression model fitted to Toys “R” Us time
series and...
FIGURE 18.19 Daily close price of Walmart stock, February 2001–
2002...
FIGURE 18.20 Department store quarterly sales series
FIGURE 18.21 Output from regression model fit to department store
sales traini...
FIGURE 18.22 Fit of regression model for department store sales
FIGURE 18.23 Monthly sales at Australian souvenir shop in dollars
(top) and in...
FIGURE 18.24 Quarterly shipments of US household appliances over
five years...
FIGURE 18.25 Monthly sales of six types of Australian wines between
1980 and 1...
Chapter 19
FIGURE 19.1 Schematic of centered moving average (top) and
trailing moving a...
FIGURE 19.2 Centered moving average (red line) and trailing moving
average (...
FIGURE 19.3 JMP time series simple moving average specification
window
FIGURE 19.4 Partial output for moving average forecaster with
applied to Am...
FIGURE 19.5 Actual vs. forecasted ridership (top) and residuals
(bottom) fro...
FIGURE 19.6 Applying MA to the residuals from the regression model
(which la...
FIGURE 19.7 The JMP Simple Exponential Smoothing dialog: change
the constrai...
FIGURE 19.8 Partial output for simple exponential smoothing
forecaster with
FIGURE 19.9 Graphs of the actual vs. forecasted ridership from the
simple ex...
FIGURE 19.10 Forecast for April 2001 residual (partial output from
saving th...
FIGURE 19.11 Partial output for Winters Method (Winters
Exponential Smoothin...
FIGURE 19.12 Graphs of actual and forecasted ridership (Top) and
residuals (...
FIGURE 19.13 State space smoothing specification window (top) and
resulting ...
FIGURE 19.14 Schematic of recurrent neural network (RNN)
FIGURE 19.15 Seasonally adjusted pre‐September 11 AIR series
FIGURE 19.16 Optimized smoothing constants
FIGURE 19.17 Optimized smoothing constants
FIGURE 19.18 Forecasts and forecast errors using regression (top)
and exponent...
FIGURE 19.19 Forecasts for validation sets
FIGURE 19.20 Quarterly shipments of US household appliances over
five years
FIGURE 19.21 Monthly sales of a certain shampoo
FIGURE 19.22 Quarterly sales of natural gas over four years
FIGURE 19.23 Monthly sales of six types of Australian wines between
1980 and 1...
Chapter 20
FIGURE 20.1 Document–term matrix representation of words for
sentences S1–S3...
FIGURE 20.2 Text Explorer dialog (top) and red triangle options
(bottom) for...
FIGURE 20.3 Document–term matrix representation of words for
modified exampl...
FIGURE 20.4 Built‐in stopwords in JMP
FIGURE 20.5 Document–term matrix after text reduction in JMP
FIGURE 20.6 TF‐IDF document–term matrix for example 2 sentences
S1–S4...
FIGURE 20.7 Importing multiple files (top) into a single JMP data
table (bot...
FIGURE 20.8 Concatenating autos.jmp and electronics.jmp to create
a combined...
FIGURE 20.9 Text explorer dialog box with specifications
FIGURE 20.10 Terms and phrase lists
FIGURE 20.11 Specifications for latent semantic analysis
FIGURE 20.12 Scatterplots from latent semantic analysis
FIGURE 20.13 Partial output from topic analysis
FIGURE 20.14 Validation metrics, confusion matrix, and lift chart for
the pr...
FIGURE 20.15 Metrics, confusion matrix, lift, and ROC curves of
validation s...
Chapter 21
FIGURE 21.1 Model accuracy (top) and AUC (bottom) for different
groups
FIGURE 21.2 False positive rates (top) and false negative rates
(bottom) for...
FIGURE 21.3 Column contributions report (COMPAS data)
FIGURE 21.4 Variable importance report (COMPAS data)
FIGURE 21.5 Marginal model plots (COMPAS data)
Chapter 22
FIGURE 22.1 RFM counts for buyers
FIGURE 22.2 RFM counts for all customers (buyers and non buyers)
MACHINE LEARNING FOR
BUSINESS ANALYTICS
GALIT SHMUELI
National Tsing Hua University
Taipei, Taiwan
PETER C. BRUCE
statistics.com
Arlington, USA
MIA L. STEPHENS
JMP Statistical Discovery LLC
Cary, USA
MURALIDHARA ANANDAMURTHY
SAS Institute Inc
Mumbai, India
NITIN R. PATEL
Cytel, Inc.
Cambridge, USA
Copyright 2023 by John Wiley & Sons, Inc. All rights reserved
Published by John Wiley & Sons, Inc., Hoboken, New Jersey.
Published simultaneously in Canada.
No part of this publication may be reproduced, stored in a retrieval system, or
transmitted in any form or by any means, electronic, mechanical, photocopying,
recording, scanning, or otherwise, except as permitted under Section 107 or 108
of the 1976 United States Copyright Act, without either the prior written
permission of the Publisher, or authorization through payment of the appropriate
per‐copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive,
Danvers, MA 01923, (978) 750‐8400, fax (978) 750‐4470, or on the web at
www.copyright.com. Requests to the Publisher for permission should be addressed
to the Permissions Department, John Wiley & Sons, Inc., 111 River Street,
Hoboken, NJ 07030, (201) 748‐6011, fax (201) 748‐6008, or online at
http://www.wiley.com/go/permission.
Trademarks: Wiley and the Wiley logo are trademarks or registered trademarks of
John Wiley & Sons, Inc. and/or its affiliates in the United States and other
countries and may not be used without written permission. All other trademarks
are the property of their respective owners. John Wiley & Sons, Inc. is not
associated with any product or vendor mentioned in this book.
Limit of Liability/Disclaimer of Warranty: While the publisher and author have used
their best efforts in preparing this book, they make no representations or
warranties with respect to the accuracy or completeness of the contents of this
book and specifically disclaim any implied warranties of merchantability or fitness
for a particular purpose. No warranty may be created or extended by sales
representatives or written sales materials. The advice and strategies contained
herein may not be suitable for your situation. You should consult with a
professional where appropriate. Neither the publisher nor author shall be liable for
any loss of profit or any other commercial damages, including but not limited to
special, incidental, consequential, or other damages.
For general information on our other products and services or for technical
support, please contact our Customer Care Department within the United States at
(800) 762‐2974, outside the United States at (317) 572‐3993 or fax (317) 572‐
4002.
Wiley also publishes its books in a variety of electronic formats. Some content that
appears in print may not be available in electronic formats. For more information
about Wiley products, visit our web site at www.wiley.com.
Library of Congress Cataloging‐in‐Publication Data Applied for:
Hardback: 9781119903833
Another random document with
no related content on Scribd:
some extraordinary rhyme which seemed as difficult to match as
Cinderella’s glass slipper, now an allusion to the events of the day,
now the sudden falling into slang in the midst of some high-flown
language.
“But Cupid is a downy cove
Wot it takes a deal to hinder;
For if you shuts him out o’ the door,
Vy, he’ll valk in at the vinder.”
Shylock. Heavy the day that first the sun the light off!
Offer rewards! Use every means to save her.
Let be but catch—I’ll lather the young shaver.
My only heiress folks will say in mock,
Fled like the timid hair from a Shy lock.
Take with you, though, unthinking girl, my curse.
Shylock. You cannot mean she’s robbed her poor old father?
This is what the fast young men of London call brilliant writing. All
this meaningless clatter of words, to produce which requires little
more skill than to clash the cymbals in the orchestra, there are
crowds of young fellows about the theatres who would give a great
deal if they had the brains to emulate. It is out of such slender
materials that Robson works up his effects, making the glitter pass
for gold, the trash for truth, the bad grammar for good sense, and the
abortive pun for pointed wit. Give us good puns by all means, if there
is nothing better to be had, and we shall laugh at them; but save us
from word-torture as incomprehensible, dull, and valueless as the
anagrams which used to puzzle and amuse our ancestors.
Cease your funning;
Force of punning
Ne’er shall Maga’s laugh trepan;
Terribly severe is all this—terrible for its truth. Gifford was not the
man to write mincingly. Nor ought we, at the present day, to write
mincingly of the iniquities and stupidities of the stage. But the fact is,
that whatever be the shortcomings of the British stage at the present
moment, and however much it may deserve the denunciations of
criticism, it is incumbent on us to dwell on those indications of
promise which are too much overlooked, rather than on the
enormous deficiencies which are patent to every observer. Let us see
whether the illustrative tendency of the time may not have its bright
side as well as a dark, and may not have a higher purpose than
spectacular effect.
It would indeed be a great mistake to imagine, that in the
production of King Henry VIII., and the other dramas that went
before it, the principal object of Mr Charles Kean was simply to place
upon the stage a dazzling spectacle, and that his success as a
manager has been due to a correct appreciation of the public taste in
this matter. Were this the case, there would be nothing special in his
managerial career. Brilliant spectacle is nothing new in the history of
the theatre—and the history of the English theatre. In the days of
James I., some of the stage properties were so very splendid, that we
have read of certain lieges who were afraid lest the double-gilt
magnificence of the tragedy-kings should cast the majesty of the real
sovereign into shade, and so endanger the crown. However absurd
and chimerical, what could be more gorgeous than the masques and
pageants which were so common in those days? Our extravaganzas
(counterparts, to a certain extent, of the ancient masque), although
they are more appropriate in costume, and altogether more matter of
fact, are not nearly so garish. Where, nowadays, shall we find a
queen willing to act like Queen Anne of Denmark—she and the ladies
of her court acting the negresses in Ben Jonson’s masque of
Blackness? Such magnificence Mr Charles Kean assuredly cannot
rival, and his claim to originality is not founded on the gorgeousness
of the spectacle which he has placed before the footlights: he claims
the praise of historical accuracy. It will be remembered how, in the
playbill of his Macbeth—a curiosity in its way—he cited the authority
of Diodorus Siculus, Pliny, Strabo, Xiphilin, Snorre, Ducange, and
the Eyrbiggia Saga—(not bad for a playbill, the Eyrbiggia Saga!)—and
in the not less remarkable programme of Sardanapalus, he lays so
great a stress on the virtues of antiquarian research and historical
fidelity, as not only to speak of his having learnt that scenic
illustration, if it have the weight of authority, may adorn and add
dignity to the noble works of genius; and to assert that in decoration
of every kind, whether scenic or otherwise, he has, in the first
instance, aimed at truth, with the grand object of conveying to the
stage an accurate portraiture and a living picture of a bygone age; but
also to point it out as a note-worthy fact that, until the present
moment, it has been impossible to render Lord Byron’s tragedy of
Sardanapalus upon the stage with proper dramatic effect, because
until now we have known nothing of Assyrian architecture and
costume; so that, according to this view, it is not enough to have for
such plays an architecture and costume artistically correct—they
must also be historically genuine. This magnifying of historical truth,
this drifting from the open and trackless sea of fiction to the terra
firma and unalterable landmarks of fact—a strong tendency to
REALISM, is the chief characteristic of Mr Kean’s management. And it
is observable not merely in his mode of placing a drama upon the
stage, but in his own style of acting. Look at Louis XI.—look at
Cardinal Wolsey, remarkable for the specification of little traits and
details that serve to realise the character as much as possible in that
style which has been called pre-Raphaelite.
Nor is this tendency peculiar to the management of the Princess’s
Theatre. It is manifested in various ways on nearly every stage
throughout the country, sometimes absurdly enough. A provincial
theatre announces a grand chivalric spectacle, “with seven hundred
pounds’ worth of real armour!” A New York theatre announces that
the School for Scandal will be produced with magnificent carpets,
mirrors, and genuine silver plate! Whittington and his Cat is
produced with a real rat amongst the crowds of sham ones, only the
sense of reality is destroyed by the terrier that plays the cat,
forgetting his catskin and beginning to yelp. One of the City theatres,
in announcing the Hertfordshire Tragedy, set forth that the very gig
in which Thurtell drove his victim to be murdered, and the very table
on which the pork-chops were afterwards devoured, would form part
of the stage properties—being expressly engaged for this theatre. In
contrast with such inane realism, one had considerable satisfaction
in gazing on the dog which Launcelot Gobbo, in Mr Talfourd’s
travesty of Shylock, so triumphantly led about—a toy-spaniel on
wheels. It is perhaps unfair to quote in such a connection the latest
vagary of this realistic tendency—a curious bit of pre-Raphaelitism—
on the part of Messrs Tom Taylor and Charles Reade, who, intending
in the King’s Rival to produce as complete a picture as possible of the
times of Charles II.—with its wit and wantonness, courtesies,
familiarities, periwigs, Mr Pepys, and Spring Gardens—actually
brought Major Wildman on the stage, in shirt and breeches, wet and
torn, and abominably plague-stricken, all the people flying from the
unsightly wretch as from an Afrit of the horrible Kaf, or a Goul of the
bottomless pit. And so, for the sake of presenting a picture of perfect
accuracy, these authors chose to turn the theatre into a Chamber of
Horrors. And since this pre-Raphaelitism, or an antiquarianism
worse than pre-Raphaelitism, is the order of the day, we are
sometimes surprised that none of the managers has seized upon that
one of Shakespeare’s plays in which, of all others, there is room for
the display of historical ingenuity, and all the originality of research.
We allude to the Tempest, and hope they will make use of the idea,
when we point out that as, according to Mr Kean, it was impossible
to represent the Sardanapalus of Lord Byron upon the stage until
Mr Layard made his discoveries at Nineveh; so, until about fifty years
ago, when Mr Malone’s Essay on the Tempest was published, it was
impossible to produce that play adequately in any theatre. The Rev.
Joseph Hunter has attempted to identify the abode of Prospero with
Lampedusa, an island half-way between Malta and the African coast,
grounding his opinion upon this amongst other facts, that
Lampedusa furnishes the Maltese with firewood, and Prospero sends
Caliban forth to collect firewood! This, however, is but child’s-play to
the labour of Malone, who not only succeeds in identifying the island
with the Bermudas, but actually discovers the identical tempest that
gives its name to the play—“the dreadful hurricane that dispersed the
fleet of Sir George Somers and Sir Thomas Gates, in July 1609, on
their passage with a large supply of provisions and men for the infant
colony in Virginia, by which the Admiral ship, as it was called, having
those commanders on board [‘some noble creatures’], was separated
from the rest of the fleet, and wrecked on the Island of Bermuda.”
Then come the incidental phrases descriptive of the storm that
identify it with the Tempest—“Admiral ship parted from the rest of
the fleet”—“they resolved to shut up the hatches”—“take leave of each
other”—“ship struck upon a rock”—“most luckily thrown up between
two, as upright as if she had been on the stocks”—“arrived in safety
without the loss of a man”—“Bermodes”—“Isle of
Devils”—“enchanted place”—“sea-monster in shape like a
man”—“richest, pleasantest, most healthful place ever seen.” What a
splendid hit Mr Kean or Mr Phelps would make if only some possible
Mr Layard could be found who should go and excavate the cell of
Prospero! Why not? Is there not perfect truth in what Mr Charles
Mathews says:—“In France the dramatic authors have free
permission to distort history ingeniously, on condition of being gay
and witty. In England, provided we are true to history, we have free
permission to be dull and tiresome.”
Now, if some of the phrases which we have been using, have been
used correctly; if we have been right in speaking of the pre-
Raphaelitism and realism of the theatre, it will be evident that the
question as to the present state of the drama, in particular, resolves
itself into a much wider question as to the present state of art
generally. And the fact is, that the more narrowly we examine the
sister arts, the more nearly do we find that they assimilate. In the
pictorial art we find the same symptoms of disintegration and decay
as in the dramatic; in both, we find the same elements of promise.
Look at the walls of our exhibition-rooms, and behold the inanities
that figure there, contemporary with the inanities of the theatre. This
picture either displays as little action as a modern tragedy, or its
action is as spasmodic as an Adelphi melodrama. In how many of
these pictures do we find the artists compensating for bad drawing
with gaudy colour, hiding vacancy of expression in a blaze of light,
feebleness of passion in a tornado of shadows, and blundering
perspective, aerial and linear, in a mist as convenient as the clouds
by which the gods of Homer saved their heroes from the lances of the
enemy? The very faults we find in the theatre! Eternal mannerism,
staginess, mimicry, trickery, grimacing, catchwords, red lights and
blue lights, and the name of the perruquier mentioned in the
playbills in large letters! In how many pictures of naked legs in the
last Exhibition, did you not recognise the calves of the gallant
grenadier who is now fighting the battles of his country? That beard,
that turban; we think we have seen the face of that Turkish Jew in at
least fifty-seven pictures; and he so haunts us throughout the
Exhibition-rooms in a thousand intolerable disguises—his long nose
here, and his cold brown eye there, as if, after using him whole as
long as possible, the artists at length cut him into little pieces, and
made a division of his remains, that really it would be a pleasure to
know that such had been his actual fate. It is the very vice of the
stage, where we find Mr A—— (who plays the villains), or Mr B——
(who plays the enamoured young gentleman), or Mr C—— (who does
the comic), eternally playing themselves, and through every possible
transformation presenting us with the same legs and arms, and
expressive nose and cracked voice. Whether on the boards or on the
canvass, incapacity and commonplace issue in virtually the same
results. And it so happens that if one were asked what are the most
striking, the most note-worthy, or the most notorious peculiarities, at
this moment, of our picture-galleries on the one hand, or of the
theatres on the other, one must inevitably fix upon the pre-
Raphaelitism of the one, and the Revivalism of the other, and
recognise them as twins. Only it must be remembered that the pre-
Raphaelitism of the picture-galleries is but one of the forms,
although the most peculiar form, in which the tendency to realism is
manifested. It is manifested not less determinately in the
prominence given to portraiture—portraits of “men, women, and
Herveys,” portraits of dogs, portraits of horses, portraits of prize
oxen and pigs, and dead game, and black-faced ewes. The colouring
which Gibson gives to his statues is a move in the same direction.
And the tendency is symbolised and strengthened by the
photographic art which has sprung up within the last few years, and
promises, whether for good or for evil, to exercise so much influence
on every easel throughout the country.
To come to the point then: What is the meaning of all this realism?
If, with all the multiform absurdities in which it is manifested, it
must nevertheless be admitted that all or most of the symptoms of
vitality in the imitative arts are at the present moment expressed in
this manner, what is the value of it?
The fact is, that whenever this tendency to realism is manifested
with more than ordinary force (we were going to say, virulence), it is
a most critical symptom. It is distinctive of what the old physicians
would have called two separate climacterics in the history of art. It
marks the infancy and the old age of art—the rise and the fall. It is
just as in the individual man—at first in childhood, and at last in
second childhood, he worships the real, and refuses to accept what
he cannot believe in as absolutely and historically true. “But is it
true?” inquires the child; “is it a fact?” says the old man. The precise
difference between the realism of infancy and that of age is another
matter to which we shall afterwards have to refer: at present we have
only to do with their generic identity. And as the individual man is in
almost every respect a miniature of the race, so we find this generic
realism characteristic of at once the beginning and the end of art. In
the middle space it culminates towards the pure azure of the ideal.
We are not sure, however, that this doctrine as to the periods of
realism, evident as it would seem to be, will obtain the immediate
acceptance of every reader: we are not sure, because the counter-
view has more than once already been put forward—and by some of
the critics in the present century has been maintained with great
vigour, that art displays most imagination in its infancy, and that—as
at once a proof and illustration of the fact—we find its most ancient
works to be the best. While the doctrine, as commonly advanced,
seems to make this wide and sweeping generalisation, it is of course
more cautiously worded, so as to apply chiefly to poetry—epical and
lyrical: as applied to the dramatic or imitative arts, there is such a
mass of evidence against it, that it could safely be advanced only by
implication. But it is not true even with reference to the narrative
poet—call him what men will—bard, aoidos, minstrel, maker,
minnesinger, scald. For observe, that the point in dispute is not
whether the most ancient poets are the best; grant for a moment that
they are: but wherein lies their distinguishing excellence? are they
more imaginative than later ones? Nothing of the kind: the
imaginative poets belong to what a geologist would call the pleiocene
formation—a much later epoch. The elder bards are remarkable
above all things for their truthfulness, their minute observation, their
naturalness, their reality. Life, the present life in the present world,
was to them an overwhelming reality, and they had no inclination,
little need, to imagine a new world, and go and live in it. A most
wonderful imagination they certainly displayed, but they were quite
unconscious of the gift: they did not imagine, like Edmund Spenser
or John Keats, for the sake of imagining; they did not dream for the
mere pleasure of dreaming. Their pleasure in dreaming was a sub-
conscious pleasure. Truth was the grand and ostensible object; and if
the facts which they proposed to discover and describe were often
mere fancies, still they were not recognised as fancies. A mere
imagination they would have regarded as a mere lie. The so-called
facts for which, in modern phrase, they were indebted to
imagination, they professed to have received from reason, from
memory, from inspiration, from veritable supernatural vision, always
from a credible source. And here, indeed, lies the strength of the
argument which refers the origin of verse to the requirements of
memory, so that versification was in its first intention but a system of
mnemonics. Right or wrong, that theory has been endorsed by
illustrious names; and it must be admitted even by an opponent, that
the whole tone of the elder poetry speaks in its favour. There is a tone
of sincerity in the elder poets, as if they could not play with their
subject, and as if upon them all had been bestowed the gift which a
fairy is said to have bestowed upon Thomas of Ercildoune—the
tongue that could not lie, the tongue that could not feign. They never
seem to be telling tales; they are relating histories. They do not
attempt to tickle the imagination; they are committing important
and interesting facts to memory. And this also is the reason why the
rhyming chroniclers—say Robert de Brunne, or Robert of Gloucester,
who were nothing but rhymers—were nevertheless regarded as true
poets. They narrated history in numerous verse: what more did those
who were truly called poets profess to do? These latter made their
narratives more interesting, but it was not recognised that the
narratives were of a different kind. Psychological analysis had not yet
penetrated so far as to discern imagination in the true poet, and none
in the rhyming chronicler. It had not yet discovered that the office of
the poet is more than this—viz., to tell what he knows faithfully,
pleasingly, and in verse. Credibility was deemed the first virtue of the
poet, the primrose of the poetical flora. What if their world be all or
half unreal?—still they believed it to be real. As it is long before the
poor mortals who have been snatched away to Elfland discover that
all the splendour which surrounds them is but a dream, that the gold
is dross, and the diamonds glass, and the brocades worsted, and the
velvets cotton, and all unreality; even so the poets of a country
(children kidnapped from a better world) do not all at once discover
that the world they live in is wholly unreal, wholly ideal. They are, at
first, the most extreme of realists.
It thus appears that even in poetry the early period is remarkable
for its realism. The poets do not begin with sublimated fancies in the