Professional Documents
Culture Documents
(Download PDF) Business Intelligence A Managerial Perspective On Analytics 3rd Edition Sharda Solutions Manual Full Chapter
(Download PDF) Business Intelligence A Managerial Perspective On Analytics 3rd Edition Sharda Solutions Manual Full Chapter
https://testbankfan.com/product/business-intelligence-a-
managerial-perspective-on-analytics-3rd-edition-sharda-test-bank/
https://testbankfan.com/product/business-intelligence-a-
managerial-perspective-on-analytics-2nd-edition-sharda-test-bank/
https://testbankfan.com/product/business-intelligence-analytics-
and-data-science-a-managerial-perspective-4th-edition-sharda-
solutions-manual/
https://testbankfan.com/product/business-intelligence-analytics-
and-data-science-a-managerial-perspective-4th-edition-sharda-
test-bank/
Business Intelligence and Analytics Systems for
Decision Support 10th Edition Sharda Solutions Manual
https://testbankfan.com/product/business-intelligence-and-
analytics-systems-for-decision-support-10th-edition-sharda-
solutions-manual/
https://testbankfan.com/product/business-intelligence-and-
analytics-systems-for-decision-support-10th-edition-sharda-test-
bank/
https://testbankfan.com/product/international-business-a-
managerial-perspective-8th-edition-griffin-solutions-manual/
https://testbankfan.com/product/business-analytics-3rd-edition-
cam-solutions-manual/
https://testbankfan.com/product/business-analytics-3rd-edition-
evans-solutions-manual/
CHAPTER
▪ Describe text analytics and understand the need for text mining
▪ Differentiate among text analytics, text mining, and data mining
▪ Understand the different application areas for text mining
▪ Know the process of carrying out a text mining project
▪ Appreciate the different methods to introduce structure to text-based data
▪ Describe sentiment analysis
▪ Develop familiarity with popular applications of sentiment analysis
▪ Learn the common methods for sentiment analysis
▪ Become familiar with speech analytics as it relates to sentiment analysis
CHAPTER OVERVIEW
CHAPTER OUTLINE
1
Copyright © 2014 Pearson Education, Inc.
Questions for the Opening Vignette
A. WHAT WE CAN LEARN FROM THIS VIGNETTE
2
Copyright © 2014 Pearson Education, Inc.
5. Financial Markets
6. Politics
7. Government Intelligence
8. Other Interesting Areas
B. SENTIMENT ANALYSIS PROCESS
1. Step 1: Sentiment Detection
2. Step 2: N-P Polarity Classification
3. Step 3: Target Identification
4. Step 4: Collection and Aggregation
C. METHODS FOR POLARITY IDENTIFICATION
D. USING A LEXICON
E. USING A COLLECTION OF TRAINING DOCUMENTS
⧫ Technology Insights 5.2: Large Textual Data Sets for
Predictive Text Mining and Sentiment Analysis
F. IDENTIFYING SEMANTIC ORIENTATION OF SENTENCES AND
PHRASES
G. IDENTIFYING SEMANTIC ORIENTATION OF DOCUMENT
Section 5.6 Review Questions
3
Copyright © 2014 Pearson Education, Inc.
⧫ Application Case 5.8: Allegro Boosts Online Click-Thru
Rates by 500 Percent with Web Analysis
B. WEB ANALYTICS METRICS
C. WEBSITE USABILITY
D. TRAFFIC SOURCES
E. VISITOR PROFILES
F. CONVERSION STATISTICS
Section 5.9 Review Questions
4
Copyright © 2014 Pearson Education, Inc.
⧫ End of Chapter Application Case: Keeping Students on
Track with Web and Predictive Analytics
Questions for the Case
References
The chapter also gives a comprehensive overview of how search engines work and the
methods used for ranking pages in response to user queries as well as the methods for
search engine optimization. Students will enjoy learning about social analytics and social
network analysis, and can relate it to other classes they may have taken in social sciences.
5
Copyright © 2014 Pearson Education, Inc.
2. What technologies were used in building Watson (both hardware and software)?
Watson is built on the DeepQA framework. The hardware for this system
involves a massively parallel processing architecture. In terms of software,
Watson uses a variety of AI-related QA technologies, including text mining,
natural language processing, question classification and decomposition, automatic
source acquisition and evaluation, entity and relation detection, logical form
generation, and knowledge representation and reasoning.
3. What are the innovative characteristics of DeepQA architecture that made Watson
superior?
4. Why did IBM spend all that time and money to build Watson? Where is the ROI?
IBM’s goal was to advance computer science by exploring new ways for
computer technology to affect science, business, and society. The techniques IBM
developed with DeepQA and Watson are relevant in a wide variety of domains
central to IBM’s mission. For example, IBM is currently working on a version of
Watson to take on surmountable problems in healthcare and medicine. If
successful, this could give IBM a distinct competitive advantage in this important
technological application area.
Text analytics is a concept that includes information retrieval (e.g., searching and
identifying relevant documents for a given set of key terms) as well as
information extraction, data mining, and Web mining. By contrast, text mining is
primarily focused on discovering new and useful knowledge from textual data
sources. The overarching goal for both text analytics and text mining is to turn
unstructured textual data into actionable information through the application of
natural language processing (NLP) and analytics. However, text analytics is a
broader term because of its inclusion of information retrieval. You can think of
text analytics as a combination of information retrieval plus text mining.
6
Copyright © 2014 Pearson Education, Inc.
2. What is text mining? How does it differ from data mining?
Text mining as a BI is increasing because of the rapid growth in text data and
availability of sophisticated BI tools. The benefits of text mining are obvious in
the areas where very large amounts of textual data are being generated, such as
law (court orders), academic research (research articles), finance (quarterly
reports), medicine (discharge summaries), biology (molecular interactions),
technology (patent files), and marketing (customer comments).
• Topic tracking. Based on a user profile and documents that a user views,
text mining can predict other documents of interest to the user.
7
Copyright © 2014 Pearson Education, Inc.
Section 5.3 Review Questions
Text mining uses natural language processing to induce structure into the text
collection and then uses data mining algorithms such as classification, clustering,
association, and sequence discovery to extract knowledge from it.
NLP moves beyond syntax-driven text manipulation (which is often called “word
counting”) to a true understanding and processing of natural language that
considers grammatical and semantic constraints as well as the context. The
challenges include:
• Word sense disambiguation. Many words have more than one meaning.
Selecting the meaning that makes the most sense can only be
accomplished by taking into account the context within which the word is
used.
8
Copyright © 2014 Pearson Education, Inc.
• Speech acts. A sentence can often be considered an action by the speaker.
The sentence structure alone may not contain enough information to
define this action.
• Question answering
• Automatic summarization
• Machine translation
• Speech recognition
• Text-to-speech
• Text proofing
1. List and briefly discuss some of the text mining applications in marketing.
Text mining can be used to increase cross-selling and up-selling by analyzing the
unstructured data generated by call centers.
9
Copyright © 2014 Pearson Education, Inc.
In 2007, EUROPOL developed an integrated system capable of accessing,
storing, and analyzing vast amounts of structured and unstructured data sources in
order to track transnational organized crime.
See Figure 5.6 (p. 222). Text mining entails three tasks:
2. What is the reason for normalizing word frequencies? What are the common
methods for normalizing word frequencies?
The raw indices need to be normalized in order to have a more consistent TDM
for further analysis. Common methods are log frequencies, binary frequencies,
and inverse document frequencies.
10
Copyright © 2014 Pearson Education, Inc.
4. What are the main knowledge extraction methods from corpus?
Sentiment analysis tries to answer the question, “What do people feel about a
certain topic?” by digging into opinions of many using a variety of automated
tools. It is also known as opinion mining, subjectivity analysis, and appraisal
extraction.
Sentiment analysis shares many characteristics and techniques with text mining.
However, unlike text mining, which categorizes text by conceptual taxonomies of
topics, sentiment classification generally deals with two classes (positive versus
negative), a range of polarity (e.g., star ratings for movies), or a range in strength
of opinion.
2. What are the most popular application areas for sentiment analysis? Why?
4. What are the main steps in carrying out sentiment analysis projects?
The first step when performing sentiment analysis of a text document is called
sentiment detection, during which text data is differentiated between fact and
opinion (objective vs. subjective). This is followed by negative-positive (N-P)
polarity classification, where a subjective text item is classified on a bipolar
11
Copyright © 2014 Pearson Education, Inc.
range. Following this comes target identification (identifying the person, product,
event, etc. that the sentiment is about). Finally come collection and aggregation,
in which the overall sentiment for the document is calculated based on the
calculations of sentiments of individual phrases and words from the first three
steps.
5. What are the two common methods for polarity identification? Explain.
1. What are some of the main challenges the Web poses for knowledge discovery?
2. What is Web mining? How does it differ from regular data mining or text mining?
Web mining is the discovery and analysis of interesting and useful information
from the Web and about the Web, usually through Web-based tools. Text mining
is less structured because it’s based on words instead of numeric data.
The three main areas of Web mining are Web content mining, Web structure
mining, and Web usage (or activity) mining.
4. What is Web content mining? How can it be used for competitive advantage?
12
Copyright © 2014 Pearson Education, Inc.
Web content mining refers to the extraction of useful information from Web
pages. The documents may be extracted in some machine-readable format so that
automated techniques can generate some information about the Web pages.
Collecting and mining Web content can be used for competitive intelligence
(collecting intelligence about competitors’ products, services, and customers),
which can give your organization a competitive advantage.
5. What is Web structure mining? How does it differ from Web content mining?
Web structure mining is the process of extracting useful information from the
links embedded in Web documents. By contrast, Web content mining involves
analysis of the specific textual content of web pages. So, Web structure mining is
more related to navigation through a website, whereas Web content mining is
more related to text mining and the document hierarchy of a particular web page.
1. What is a search engine? Why are they important for today’s businesses?
A search engine is a software program that searches for documents (Internet sites
or files) based on the keywords (individual words, multi-word terms, or a
complete sentence) that users have provided that have to do with the subject of
their inquiry. This is the most prominent type of information retrieval system for
finding relevant content on the Web. Search engines have become the centerpiece
of most Internet-based transactions and other activities. Because people use them
extensively to learn about products and services, it is very important for
companies to have prominent visibility on the Web; hence the major effort of
companies to enhance their search engine optimization (SEO).
A Web crawler (also called a spider or a Web spider) is a piece of software that
systematically browses (crawls through) the World Wide Web for the purpose of
finding and fetching Web pages. It starts with a list of “seed” URLs, goes to the
pages of those URLs, and then follows each page’s hyperlinks, adding them to the
search engine’s database. Thus, the Web crawler navigates through the Web in
order to construct the database of websites.
13
Copyright © 2014 Pearson Education, Inc.
primarily benefits companies with e-commerce sites by making their pages appear
toward the top of search engine lists when users query.
4. What things can help Web pages rank higher in the search engine results?
Cross-linking between pages of the same website to provide more links to the
most important pages may improve its visibility. Writing content that includes
frequently searched keyword phrases, so as to be relevant to a wide variety of
search queries, will tend to increase traffic. Updating content so as to keep search
engines crawling back frequently can give additional weight to a site. Adding
relevant keywords to a Web page’s metadata, including the title tag and
metadescription, will tend to improve the relevancy of a site’s search listings, thus
increasing traffic. URL normalization of Web pages so that they are accessible via
multiple URLs. Using canonical link elements and redirects can help make sure
links to different versions of the URL all count toward the page’s link popularity
score.
1. What are the three types of data generated through Web page visits?
14
Copyright © 2014 Pearson Education, Inc.
4. What are commonly used Web analytics metrics? What is the importance of metrics?
• Website usability: How were they using my website? These involve page
views, time on site, downloads, click map, and click paths.
• Traffic sources: Where did they come from? These include referral
websites, search engines, direct, offline campaigns, and online campaigns.
• Visitor profiles: What do my visitors look like? These include keywords,
content groupings, geography, time of day, and landing page profiles.
• Conversion statistics: What does all this mean for the business? Metrics
include new visitors, returning visitors, leads, sales/conversions, and
abandonments.
These metrics are important because they provide access to a lot of valuable
marketing data, which can be leveraged for better insights to grow your
business and better document your ROI. The insight and intelligence gained
from Web analytics can be used to effectively manage the marketing efforts of
an organization and its various products or services.
15
Copyright © 2014 Pearson Education, Inc.
3. What is social media? How does it relate to Web 2.0?
4. What is social media analytics? What are the reasons behind its increasing
popularity?
Social media analytics refers to the systematic and scientific ways to consume the
vast amount of content created by Web-based social media outlets, tools, and
techniques for the betterment of an organization’s competitiveness. Data includes
anything posted in a social media site. The increasing popularity of social media
analytics stems largely from the similarly increasing popularity of social media
together with exponential growth in the capacities of text and Web analytics
technologies.
First, determine what your social media goals are. From here, you can use
analysis tools such as descriptive analytics, social network analysis, and advanced
(predictive, text examining content in online conversations), and ultimately
prescriptive analytics tools.
16
Copyright © 2014 Pearson Education, Inc.
various data sources (patent databases, new release archives, and product
announcements) in order to develop a holistic view of the competitive landscape.
3. What were the challenges, the proposed solution, and the obtained results?
Kodak’s challenges are to apply more than a century’s worth of knowledge about
imaging science and technology to new uses and to secure those new uses with
patents. The problem is that it is nearly impossible to efficiently process such
enormous amounts of semistructured data (patent documents usually contain
partially structured and partially textual data). But through the use of text mining
tools, Kodak is able to obtain a wide range of benefits. These include enabling
competitive intelligence, making critical business decisions, identifying and
recruiting new talent, identifying unauthorized use of Kodak’s patents, identifying
complementary inventions for forming symbiotic partnerships, and preventing
competitors from creating similar products.
Application Case 5.2: Text Mining Improves Hong Kong Government’s Ability to
Anticipate and Address Public Complaints
1. How did the Hong Kong government use text mining to better serve its
constituents?
2. What were the challenges, the proposed solution, and the obtained results?
The major challenge facing the Hong Kong government was analyzing and
responding to public complaints to their call center. This involves addressing
answers to about 2.65 million calls and 98,000 e-mails per year. Originally, they
attempted to compile reports on complaint statistics for reference by government
departments manually. But through ‘eyeball’ observations, it was impossible to
effectively reveal new or more complex potential public issues and identify their
root causes, as most of the complaints were recorded in unstructured textual
format. So, the government decided to utilize text processing and mining
approaches to uncover trends, patterns, and relationships in the text data. This
allowed the government to better understand the voice of the people, improve
service delivery, make informed decisions, and develop smart strategies. The
result was a boost in public satisfaction.
17
Copyright © 2014 Pearson Education, Inc.
1. Why is it difficult to detect deception?
3. What do you think are the main challenges for such an automated system?
One challenge is that the training the system depends on humans to ascertain the
truthfulness of statements in the training data itself. You can’t know for sure
whether these statements are true or false, so you may be using incorrect training
samples when “teaching” the machine learning system to predict lies in new text
data. (This answer will vary by student.)
Application Case 5.4: Text Mining and Sentiment Analysis Help Improve Customer
Service Performance
1. How did the financial services firm use text mining and text analytics to improve
its customer service performance?
The company used PolyAnalyst’s text analysis tools for extracting complex word
patterns, grammatical and semantic relationships, and expressions of sentiment.
The results were classified into context-specific themes to identify actionable
issues. The relationships between structured fields and text analysis results were
established to identify patterns and interactions. Results were presented via
graphical, interactive, web-based reports. Actionable issues were assigned to
relevant individuals responsible for their resolution, and criteria were established
to capture and detect compliance with the company’s Quality Standards.
2. What were the challenges, the proposed solution, and the obtained results?
18
Copyright © 2014 Pearson Education, Inc.
Continually monitoring service levels is essential for service quality control.
Customer surveys and associate-customer interactions are the best way to obtain
data for such monitoring. But manually evaluating this data is subjective, error-
prone, and time/labor-intensive. Therefore, the company needed a system for (1)
automatically evaluating associate-customer interactions for compliance with
quality standards and (2) analyzing survey responses to extract positive and
negative feedback, while allowing for the diversity of natural language
expression. The solution was to use PolyAnalyst. (See answer to Question #1 for
the remainder of the answer.)
1. How can text mining be used to ease the insurmountable task of literature review?
2. What are the common outcomes of a text mining project on a specific collection
of journal articles? Can you think of other potential outcomes not mentioned in
this case?
Application Case 5.6: Whirlpool Achieves Customer Loyalty and Product Success
with Text Analytics
1. How did Whirlpool use capabilities of text analytics to better understand their
customers and improve product offerings?
Whirlpool uses Attensity products for deep text analytics of their multi-channel
customer data, which includes e-mails, CRM notes, repair notes, warranty data,
and social media. The company uses text analytics solutions every day to get to
the root cause of product issues and receive alerts on emerging issues. Users of
Attensity’s analytics products at Whirlpool include product/ service managers,
19
Copyright © 2014 Pearson Education, Inc.
corporate/product safety staff, consumer advocates, service quality staff,
innovation managers, the Category Insights team, and all of Whirlpool’s
manufacturing divisions (across five countries).
2. What were the challenges, the proposed solution, and the obtained results?
Customer satisfaction and feedback are at the center of how Whirlpool drives its
overarching business strategy, so gaining insight into customer satisfaction and
product feedback is paramount. Whirlpool needed to more effectively understand
and react to customer and product feedback data, originating from blogs, e-mails,
reviews, forums, repair notes, and other data sources. Managers needed to report
on longitudinal data, and be able to compare issues by brand over time. The
solution was to use Attensity’s Text Analytics application. This enabled the
company to more proactively identify and mitigate quality issues before issues
escalated and claims were filed. Since then, Whirlpool has also been able to avoid
recalls, which increased customer loyalty and reduced costs (realizing 80%
savings on their costs of recalls due to early detection).
Lotte.com introduced its integrated Web traffic analysis system using the SAS for
Customer Experience Analytics solution. This enabled Lotte.com to accurately
measure and analyze website visitor numbers (UV), page view (PV) status of site
visitors and purchasers, the popularity of each product category and product,
clicking preferences for each page, the effectiveness of campaigns, and much
more. This information enables Lotte.com to better understand customers and
their behavior online, and conduct sophisticated, cost-effective targeted
marketing.
2. What were the challenges, the proposed solution, and the obtained results?
With almost 1 million website visitors each day, Lotte.com needed to know how
many visitors were making purchases and which channels were bringing the most
valuable traffic. SAS for Customer Experience Analytics solution fully
transformed the Lotte.com website. As a result, Lotte.com has been able to
improve the online experience for its customers as well as generate better returns
from its marketing campaigns. This led to a jump in customer loyalty, improved
market analysis, a jump in customer satisfaction, and higher sales.
20
Copyright © 2014 Pearson Education, Inc.
To the degree that e-commerce companies integrate analytics into their systems,
they can take advantage of these technologies to make improvements in the user
experience on their e-commerce systems and thereby increase customer
satisfaction.
Application Case 5.8: Allegro Boosts Online Click-Thru Rates by 500 Percent with
Web Analysis
1. How did Allegro significantly improve click-through rates with Web analytics?
2. What were the challenges, the proposed solution, and the obtained results?
Allegro’s challenge was how to match the right offer to the right customer while
still being able to support the extraordinary amount of data it held. The company
had over 75 proprietary websites in 11 European countries, over 15 million
products, and generated over 500 million page views per day. In particular
Allegro aimed to increase its income and gross merchandise volume from its
current network, as measured by click-through rates and conversion rates. The
solution was to implement the SNA-based product recommendation system
described above. The results were dramatic. User traffic increased by 30%, click-
through rates increased by 500%, and conversion rates are 40 times what they
were previously.
It can help telecommunications companies listen to and understand the needs and
wants of customers. This enables them to offer communication plans, prices, and
features that are tailored to the individual customer. SNA can be applied to call
records that already exist in a telecommunication company’s database. This data
is analyzed using SNA metrics such as influencers, degree, density, betweenness
and centrality.
2. What do you think are the key challenges, potential solution, and probable results
in applying SNA in telecommunications firms?
21
Copyright © 2014 Pearson Education, Inc.
Because of the widespread use of free Internet tools and techniques (VoIP, video
conferencing tools such as Skype, free phone calls within the United States by
Google Voice, etc.), the telecommunication industry is going through a tough
time. In order to stay viable and competitive, companies need to make the right
decisions and utilize their limited resources optimally. Market pressures force
telecommunication companies to be more innovative. Using SNA, companies can
manage churn, improve cross-sell and technology transfer, manage viral
campaigns, better identify customers, and gain competitor insight.
1. How did C3 Presents use social media analytics to improve its business?
The company contracted with Cardinal Path to architect and implement a solution
based on an existing Google Analytics implementation. A combination of
customized event tracking, campaign tagging, custom variables, and a complex
implementation and configuration was deployed to include the tracking of each
social media outlet on the site.
2. What were the challenges, the proposed solution, and the obtained results?
Application Case 5.11: eHarmony Uses Social Media to Help Take the Mystery Out
of Online Dating
22
Copyright © 2014 Pearson Education, Inc.
mining that identifies and classifies terms in text sources according to sentiment
(e.g. judgment, opinion, and emotional content).
2. In your own words, define text mining and discuss its most popular applications.
3. What does it mean to induce structure into the text-based data? Discuss the
alternative ways of inducing structure into text-based data.
Text mining, like other data mining approaches, is an inductive approach for
finding patterns and trends in data. One difference between text mining and other
data mining approaches is the use of natural language processing.
4. What is the role of natural language processing in text mining? Discuss the
capabilities and limitations of NLP in the context of text mining.
5. List and discuss three prominent application areas for text mining. What is the
common theme among the three application areas you chose?
23
Copyright © 2014 Pearson Education, Inc.
Marketing: Text mining can be used to increase cross-selling and up-selling by
analyzing the unstructured data generated by call centers.
Sentiment analysis tries to answer the question, “What do people feel about a
certain topic?” by digging into opinions of many using a variety of automated
tools. It is also known as opinion mining, subjectivity analysis, and appraisal
extraction. Sentiment analysis shares many characteristics and techniques with
text mining. However, unlike text mining, which categorizes text by conceptual
taxonomies of topics, sentiment classification generally deals with two classes
(positive versus negative), a range of polarity (e.g., star ratings for movies), or a
range in strength of opinion.
7. What are the common challenges that sentiment analysis has to deal with?
Sentiment that appears in text comes in two flavors: explicit, where the subjective
sentence directly expresses an opinion (“It’s a wonderful day”), and implicit,
where the text implies an opinion (“The handle breaks too easily”). Implicit
sentiment analysis is harder to analyze because it may not include words that are
obviously evaluations or judgments. Another challenge involves the timeliness of
collection/analysis of textual data coming from a wide variety of data sources. A
third challenge is the difficulty of identifying whether a piece of text involves
sentiment or not, especially with implicit sentiment analysis. The same sorts of
issues involving text mining in natural language settings also apply to sentiment
analysis.
8. What are the most popular application areas for sentiment analysis? Why?
9. What are the main steps in carrying out sentiment analysis projects?
The first step when performing sentiment analysis of a text document is called
sentiment detection, during which text data is differentiated between fact and
opinion (objective vs. subjective). This is followed by negative-positive (N-P)
polarity classification, where a subjective text item is classified on a bipolar
range. Following this comes target identification (identifying the person, product,
event, etc. that the sentiment is about). Finally come collection and aggregation,
in which the overall sentiment for the document is calculated based on the
calculations of sentiments of individual phrases and words from the first three
steps.
24
Copyright © 2014 Pearson Education, Inc.
10. What are the two common methods for polarity identification? Explain.
11. Discuss the differences and commonalities between text mining and Web mining.
12. In your own words, define Web mining and discuss its importance.
13. What are the three main areas of Web mining? Discuss the differences and
commonalities among these three areas.
Web mining consists of three areas: Web content mining, Web structure mining,
and Web usage mining.
Web content mining refers to the automatic extraction of useful information from
Web pages. It may be used to enhance search results produced by search engines.
Web structure mining refers to generating interesting information from the links
included in Web pages. Web structure mining can also be used to identify the
members of a specific community and perhaps even the roles of the members in
the community.
14. What is a search engine? Why are they important for businesses?
A search engine is a software program that searches for documents (Internet sites
or files) based on the keywords (individual words, multi-word terms, or a
complete sentence) that users have provided that have to do with the subject of
their inquiry. This is the most prominent type of information retrieval system for
finding relevant content on the Web. Search engines have become the centerpiece
of most Internet-based transactions and other activities. Because people use them
25
Copyright © 2014 Pearson Education, Inc.
extensively to learn about products and services, it is very important for
companies to have prominent visibility on the Web; hence the major effort of
companies to enhance their search engine optimization (SEO).
15. What is search engine optimization? Who benefits from it? How?
16. What is Web analytics? What are the metrics used in Web analytics?
Web analytics, also called Web usage mining, can be considered a part of Web
mining, and aims to describe what has happened on the website (employing a
predefined, metrics-driven descriptive analytics methodology). There are four
main categories of Web analytic metrics:
• Website usability: How were they using my website? These involve page
views, time on site, downloads, click map, and click paths.
• Traffic sources: Where did they come from? These include referral
websites, search engines, direct, offline campaigns, and online campaigns.
• Visitor profiles: What do my visitors look like? These include keywords,
content groupings, geography, time of day, and landing page profiles.
• Conversion statistics: What does all this mean for the business? Metrics
include new visitors, returning visitors, leads, sales/conversions, and
abandonments.
These metrics are important because they provide access to a lot of valuable
marketing data, which can be leveraged for better insights to grow your business
and better document your ROI. The insight and intelligence gained from Web
analytics can be used to effectively manage the marketing efforts of an
organization and its various products or services.
17. Define social analytics, social network, and social network analysis. What are the
relationships among them?
26
Copyright © 2014 Pearson Education, Inc.
a social structure composed of individuals/people (or groups of individuals or
organizations) linked to one another with some type of connections/relationships.
Social network analysis (SNA) is the systematic examination of social networks,
and is an interdisciplinary field that emerged from social psychology, sociology,
statistics, and graph (network) theory. Social analytics therefore combines text
analysis for content and sentiment in online communications with social network
analysis to identify and analyze relationships between individuals in a
community.
18. What is social media analytics? How is it done? Who does it? What comes out of
it?
Social media analytics refers to the systematic and scientific ways to consume the
vast amount of content created by Web-based social media outlets, tools, and
techniques for the betterment of an organization’s competitiveness. It is done
using many analytic methods, including text mining, sentiment analysis, and
social network analysis. Companies use it to get a better understanding of their
customer base, and can gain financial and competitive advantages from doing so.
Governments use it to track potential terrorist threats, which can lead to enhanced
national security. Social scientists use it to get a better understanding of how
communities and societies work, which can provide guidance on how to best
manage these societies.
One big challenge is keeping and improving retention in an online university. But
educators lacked the hard data to prove conclusively what really drives retention.
Colleges and universities have struggled to understand how to lower dropout rates
and keep students on track all the way to graduation. APUS faced this challenge in a
new online environment, which further complicates the retention dilemma.
2. What types of data did APUS tap into? What do you think are the main obstacles one
would have to overcome when using data that comes from different domains and
sources?
APUS tapped into demographic data, registration data, and course level data. They
used online activity as well as end-of-course survey data. (The remainder of this
answer may vary according to student.) One obstacle with bringing in data from
different domains is the complexity of finding meaningful patterns in the data. This
could require data extraction, cleaning, and transforming the source data.
27
Copyright © 2014 Pearson Education, Inc.
3. What solutions did they pursue? What tools and technologies did they use?
They used predictive analytics to zero in on those factors that most influence a
student’s decision to stay in school or drop out. In addition, they built dashboards to
put the predictive analytics into the hands of deans and other administrators.
4. What were the results? Do you think these results are also applicable to other
educational institutions? Why? Why not?
Some of the findings generated by its predictive models actually came as a surprise to
APUS. For example, it had long been assumed that gender and ethnicity were good
predictors of attrition, but the models proved otherwise. Educators also assumed that
a preparatory course called “College 100: Foundations of Online Learning” was a
major driver of retention, but an in-depth analysis using Modeler came to a different
conclusion. Also, the results showed that a sense of community was an important
retention factor. So, APUS is putting more effort into promoting interaction between
students via social media and online collaboration tools.
Predictive analytics are clearly applicable at other universities. But some metrics may
not apply. For example, click-stream analysis won’t apply for universities with
traditional face-to-face classroom delivery.
5. What additional analysis are they planning on conducting? Can you think of other
data analyses that they could apply and benefit from?
APUS also plans to delve deeper into the surveys by mining the open-ended text
responses that are part of each questionnaire (IBM SPSS Text Analytics for Surveys
will help with that initiative). All of these data-driven initiatives aim to increase
student learning, enhance the student’s experience, and build an environment that
encourages retention at APUS.
28
Copyright © 2014 Pearson Education, Inc.
Another random document with
no related content on Scribd:
sydämeensä. Hän tahtoo varoittaa ja tuntee voivansa itsekin vielä
olla varoituksen tarpeessa.
Eduard Charlottalle.
*****