Download as pdf or txt
Download as pdf or txt
You are on page 1of 41

ASSIGNMENT 1 FRONT SHEET

Qualification BTEC Level 5 HND Diploma in Business

Unit number and title Unit 31: Statistics for management

Submission date Date Received 1st submission

Re-submission Date Date Received 2nd submission

Student Name Phan Huu Dai Anh Student ID GBS190316

Class GBS0714_PPT Assessor name Nguyen Thi Bich Diep

Student declaration

I certify that the assignment submission is entirely my own work and I fully understand the consequences of plagiarism. I understand that
making a false declaration is a form of malpractice.

Student’s signature

Grading grid

P1 P2 M1
 Summative Feedback:  Resubmission Feedback:

Grade: Assessor Signature: Date:


Internal Verifier’s Comments:

Signature & Date:


Assignment Brief 1 (RQF)

Higher National Certificate/Diploma in Business

Unit Number and Title Unit 31 - Statistics for Management

Academic Year 2020-2021

Unit Assessor

Assignment Title ASSIGNMENT 1 – Statistical Methods

Submission Date

Submission Format

Format

 This assignment is an Individual report.


 You must use font Times New Roman, size 12, set number of the pages and line spacing at 1.5
lines. Margins must be: left: 2.5 cm; right: 2 cm; top: 2 cm and bottom: 2 cm.
 You should use in-text references and a list of all cited sources at the end of the essay by
applying the Harvard referencing style.
 The recommended word limit is 1800-2000 words (+/-10%), excluding the tables, graphs,
diagrams, appendixes and references. You will not be penalized for exceeding the total word
limit.
 The cover page of the report has to be the Assignment front sheet 1 (to be attached with this
assignment brief).

Submission

 Students are compulsory to submit the assignment in due date (slot 12) and in a way
requested by the Tutor.
 The form of submission will be a soft copy posted on http://cms.greenwich.edu.vn/.
 Remember to convert the word file into PDF file before the submission on CMS.

Note

 The individual Assignment must be your own work, and not copied by or from another
student.
 If you use ideas, quotes or data (such as diagrams) from books, journals or other sources, you
must cite your sources, using the Harvard style.
 Make sure that you understand and follow the guidelines to avoid plagiarism. Failure to
comply with this requirement will result in a failed assignment.

Unit Learning Outcomes

LO1. Evaluate business and economic data/information obtained from published sources.

Assignment Brief and Guidance

*This assignment guidance is for reference only and can be customized by the tutor to meet specific needs

Assignment Scenario

You are assumed as a Research Analyst. Your company is planning to improve the information
system and the decision-making process by applying several statistical methods.

Precisely, you are required to demonstrate your understanding by evaluating and analysing the
business data (financial information, stock market) or microeconomics or macroeconomic recent
problems, future trends/ intentions, etc. regarding your research topic. All the variables could be
nominal or ordinal, interval, or ratio. Besides, you have to make sure that your dataset characterised
by the rules: Composing of at least 6 variables and 50 observations to do the research.

In this assignment, you have to complete the following requirements:

Introduction

1. Background and the reasons why you choose the topic


2. Objectives, scope, and meaning of the study
3. Methodology
4. Structure of the report

Main contents

1. Evaluate the nature and process of business and economic data/information from a range of
different published sources
 Define data, information and knowledge.
 From published sources and theoretical frameworks of the primary research, you should show
how data can be turned into information and information into knowledge.
 Describe in detail the way of collecting database and transformation process of data into
information and knowledge.
2. Evaluate data from a variety of sources using different methods of analysis
 Interpreting data from a variety of sources using different methods of analysis:
descriptive, exploratory and confirmatory.
 Demonstrating the pros and cons of each method about your above interpretation.
 Critically evaluating the differences in application among descriptive, exploratory and
confirmatory analysis techniques of business and economic data in general.

Unit Assessment Criteria

Pass Merit
LO1. Evaluate business and economic data/information
obtained from published sources

P1 Evaluate the nature and


process of business and
economic data/information M1 Critically evaluate the
from a range of different methods of analysis used to
published sources. present business and economic
data/information from a range
P2 Evaluate data from a variety of different published sources.
of sources using different
methods of analysis.
Table of content
I. Introduction....................................................................................................................................... 7

1. Background and the reasons why I choose the topic.............................................................. 7

2. Objectives, scope, and meaning of the study.......................................................................... 7

3. Methodology............................................................................................................................. 8

4. Structure of the report.............................................................................................................. 8

II. Evaluate the nature and process of business and economic data/information............................8

1. Definition of data, information and knowledge...................................................................... 8

2. How data can be turned into information and information into knowledge.......................10

3. The way of collecting database and transformation process of data into information and
knowledge................................................................................................................................... 16

III. Evaluate data from a variety of sources using different methods of analysis............................21

*Descriptive analysis................................................................................................................... 21

*Exploratory analysis.................................................................................................................. 25

*Confirmatory analysis............................................................................................................... 27

*Demonstrating the pros and cons of each method about your above interpretation...........30

IV. The analysis of the research of the behaviour of people who are dealing with COVID-19 in
America............................................................................................................................................... 32

V. Conclusion and suggestion.............................................................................................................39

Reference list:..................................................................................................................................... 40
I. Introduction

1. Background and the reasons why I choose the topic

Since the COVID-19 pandemic began, Ipsos has partnered with Axios to release a weekly
Coronavirus Index survey. The survey examines Americans' behaviors, perceptions of risk, and
attitudes around COVID-19.

Ipsos Group S.A. is a multinational market research and consulting firm with headquarters
in Paris, France. The company was founded in 1975 by Didier Truchot, Chairman and CEO, and has
been publicly traded on the Paris Stock Exchange since July 1, 1999.

Besides, Axios is an American news website based in Arlington County, Virginia. It was
founded in 2016 and launched the following year by former Politico journalists Jim VandeHei, Mike
Allen and Roy Schwartz. The site's name is based on the Greek: ἄξιος (áxios), meaning "worthy".

The reason why I choose this topic is because the influence of Covid 19 globally has
greatly affected the general economies of many countries. In particular, if we study and show how
well the behavior of the current American population against COVID-19 will help revive the
economy after the epidemic is under control.

2. Objectives, scope, and meaning of the study

 Objectives:

The goal is to study the behavior of the American people in the emerging COVID-19
situation. To make plans for economic recovery and projections for open door policies after the
epidemic has been controlled.

 Scope of Project:

Geographic Coverage: View help for Geographic Coverage U.S., general population age
18+

Time Period(s): View help for Time Period(s) 3/2020 – 12/2020 (Research is ongoing, but
the data uploaded (waves 1-33) covers March through December, 2020.)
 Meaning of the study:

This study will show the behavior of people in the face of a major pandemic. From there,
the government will create new policies to prevent the situation when a new pandemic emerges.

3. Methodology

The methodology used in this study is Quantitative - Questionnaire. This method will
help the research to be investigated in a wide range across the United States. But in this report, the
results will be reduced to 151 people compared to the official results and there are only 6 different
variables in this report.

4. Structure of the report

+ Introduction

+ Main body:

1.Evaluate the nature and process of business and economic data/information

2.Evaluate data from a variety of sources using different methods of analysis.

+ Conclusion and Suggestion

II. Evaluate the nature and process of business and economic data/information

1. Definition of data, information and knowledge

*Data

According to BYJU (2020), data is the complete list of facts and details like text,
observations, figures, symbols and description of things. It is the raw list of facts that are processed
to gain information. The basic concept of data is associated with scientific research collected by
different research organizations.

For example, the information collected for writing a research paper is data until it is
presented in an organized manner.
Or others example of data could be like the below table

(Source: USDA Ag Data Commons)

*Information

Information is the processed, organized data that is beneficial in providing useful


knowledge. For example, the data compiled in an organized way in a research paper provides
information about a particular concept. (BYJU, 2020)
*Knowledge

The next level is knowledge – this step gravitates towards the How question. As a pool of
in-context information has been created over various random periods of time, knowledge is the
output of experience that are interpreted and then generated from that collection of identified
information.

For example:

2. How data can be turned into information and information into knowledge.

This section will use the DIKW framework to demonstrate how data can be turned into
information and information into knowledge. The DIKW pyramid, also known variously as the
DIKW hierarchy, wisdom hierarchy, knowledge hierarchy, information hierarchy, and the data
pyramid, refers loosely to a class of models for representing purported structural and/or functional
relationships between data, information, knowledge, and wisdom. "Typically information is defined
in terms of data, knowledge in terms of information, and wisdom in terms of knowledge".
(CertGuidance, 2020)
Not all versions of the DIKW model reference all four components (earlier versions not
including data, later versions omitting or downplaying wisdom), and some include additional
components. In addition to a hierarchy and a pyramid, the DIKW model has also been characterized
as a chain, as a framework, as a series of graphs, and as a continuum. (CertGuidance, 2020)

The DIKW model of transforming data into wisdom can be viewed from two different
angles: contextual and understanding. As per the contextual concept, one moves from a phase of
gathering data parts (data), the connection of raw data parts (information), formation of whole
meaningful contents (knowledge) and conceptualize and joining those whole meaningful contents
(wisdom). From the understanding perspective, the DIKW Pyramid can also be seen as a process
starting with researching & absorbing, doing, interacting, and reflecting. The DIKW hierarchy can
also be represented in terms of time. For instance, the data, information, and knowledge levels can
be seen as the past while the final step - wisdom - represents the future.

(Source: CertGuidance.com)
+ Level 1: Data

- Purely it is just "raw" data (can be collected from before or via sensors that transmit data
in real time)

- This form is not common on ordinary reading sources such as books, magazines (because
books and magazines are required to process data into higher formats in order to attract and attract
readers)

- Often seen when we need to study a problem (related to data, tables, ...)

- The truths are obvious, the truths, can be proved with evidence.

- Can be considered a kind of "information" at a low level.

+ Level 2: Information = Data + context

#Understanding Relations --- #Physical structuring

- From this huge data warehouse, by sorting, reorganizing them in groups, in order, by
relationships - Relation

(cause - effect, before - after) and placed in context - specific Context for which we get
"Information" - "useful" data form

- Often found in books, entertainment, ... to serve the needs of exchanging and processing
information of people, is the provision of information about Who, what, when and where - Who ?
What? When? Where?

- Connect to the Data layer at the primary level - Direct Connect (Physical)

Of course information exists in both forms: True and False (even deliberately false - Fake
News)

- When presenting this information to others, it is merely "repeat / forward"


+ Level 3: Knowledge = Information + meaning

#Understanding Patterns --- #Cognitive structuring

- Going up the next ladder is to level 3: Knowledge

- It is the Information that is accompanied by their Meaning - Meaning

- Is the result after a long enough time of studying, discussing, receiving information
processing, ideas, concepts, as well as comparing, comparing ... to find out the correct, useful
information. Building knowledge Foundation!

-> False Knowledge needs to be eliminated by Correct Knowledge.

- Going to answer 2 more difficult questions than: How? Suppose? - How? What if?

-> Learn about Patterns, motifs, patterns - Patterns, to get a quantity of input information.
After the research process, think for knowledge and understanding at a certain level about a topic,
field or industry!

- Connect with the Information layer at a higher level - Connect at the level of Cognitive
(Cognitive)

- Gradually turn outside information and knowledge of others into your own knowledge, so
that when re-presenting that knowledge, it will be in the form of your own Language.
(Source: CertGuidance.com)

+ Level 4: Wisdom = Knowledge + insight

#Understanding Principles --- #Belief structuring

- Is the highest floor of the DIKW Tower: Thinking - Wisdom - Wisdom!

- This is the greatest asset, the privilege that only Humans have - Is a collection of a full
amount of Knowledge in a certain field, which is continuously upgraded, deepened, applied,
criticized, analysis, ... to lead to understanding - Insight

- Insight: It is the implicit truths, the insight(surely everyone already knows the phrase:
Business insight, Customer insight)
-> 1 person may be very intelligent in one industry but zero in another! -> DIKW Tower
can completely explain this!

- Learn at the highest level: Principle, Essence, Core!

(Relation -> Pattern -> Principle), delving into the very essence of the matter, what is the
root cause, the root cause of all this, answers the hardest question: Why ? - Why?

- Connect with Knowledge at the highest level - Connect with Belief (Belief)

-> At this point, all knowledge is synthesized, merged into one body, they have a
panoramic view, insight, details as well as confidence in presenting. In addition, they will debate
almost every aspect around their field and be considered "experts" in that area!

(Source: Guru99)
3. The way of collecting database and transformation process of data into
information and knowledge.

*Data collection process

According to Ainsworth (2020), although data can be valuable, too much information is
unwieldy, and the wrong data is useless. The right data collection method can mean the differences
between useful insights and time-wasting misdirection.

There 6 data collection tools:

+ Interviews

If you asked someone completely unaware of data analysis how to best collect information
from people, the most common answer would likely be interviews. Almost anyone can come up
with a list of questions, but the key to efficient interviews is knowing what to ask. Efficiency in
interviewing is crucial because, of all the primary data collection methods, in-person interviewing
can be the most expensive. ( Ainsworth, 2020)

+ Observation

Observation involves collecting information without asking questions. This method is more
subjective, as it requires the researcher, or observer, to add their judgment to the data. But in some
circumstances, the risk of bias is minimal. ( Ainsworth, 2020)

For example, if a study involves the number of people in a restaurant at a given time, unless
the observer counts incorrectly, the data should be reasonably reliable. Variables that require the
observer to make distinctions, such as how many millennials visit a restaurant in a given period, can
introduce potential problems. In general, observation can determine the dynamics of a situation,
which generally cannot be measured through other data collection techniques. Observation also can
be combined with additional information, such as video. ( Ainsworth, 2020)
+ Documents and records

Sometimes you can collect a considerable amount of data without asking anyone anything.
Document- and records-based research uses existing data for a study. Attendance records, meeting
minutes, and financial records are just a few examples of this type of research. ( Ainsworth, 2020)

Using documents and records can be efficient and inexpensive because you’re
predominantly using research that has already been completed. However, since the researcher has
less control over the results, documents and records can be an incomplete data source. ( Ainsworth,
2020)

+ Focus groups

A combination of interviewing, surveying, and observing, a focus group is a data collection


method that involves several individuals who have something in common. The purpose of a focus
group is to add a collective element to individual data collection. A focus group study can ask
participants to watch a presentation, for example, then discuss the content before answering survey
or interview-style questions. ( Ainsworth, 2020)

Focus groups often use open-ended questions such as, “How did you feel about the
presentation?” or “What did you like best about the product?” The focus group moderator can ask
the group to think back to the shared experience, rather than forward to the future. Open-ended
questions ground the research in a particular state of mind, eliminating external interference.
( Ainsworth, 2020)

+ Oral histories

At first glance, an oral history might sound like an interview. Both data collection methods
involve asking questions. But an oral history is more precisely defined as the recording, preservation,
and interpretation of historical information based on the opinions and personal experiences of people
who were involved in the events. Unlike interviews and surveys, oral histories are linked to a single
phenomenon. For example, a researcher may be interested in studying the effect of a flood on a
community. An oral history can shed light on exactly what transpired. It’s a holistic approach to
evaluation that uses a variety of techniques. ( Ainsworth, 2020)
As in interviewing, the researcher can become a confounding variable. A confounding
variable is an extra, unintended variable that can skew your results by introducing bias and
suggesting a correlation where there isn’t one. The classic example is the correlation between
murder rates and ice cream sales. Both figures have, at one time or another, risen together. An
unscientific conclusion may be that the more people buy ice cream, the higher the occurrence of
murder. However, there is a third possibility that an additional variable affects both of these
occurrences. In the case of ice cream and murder, the other variable is the weather. Warmer weather
is a confounding variable to both murder rates and ice cream sales. ( Ainsworth, 2020)

+ Questionnaires and surveys

Questionnaires and surveys can be used to ask questions that have closed-ended answers.
Data gathered from questionnaires and surveys can be analyzed in many different ways. You can
assign numerical values to the data to speed up the analysis. This can be useful if you’re collecting a
large amount of data from a large population. ( Ainsworth, 2020)

To be meaningful, surveys and questionnaires need to be carefully planned. Unlike an


interview, where a researcher can react to the direction of a respondent’s answers, a poorly designed
questionnaire will lead the study nowhere quickly. While surveys are often less expensive than
interviews, they won’t be valuable if they aren’t handled correctly. ( Ainsworth, 2020)

Surveys can be conducted as interviews, but in most cases, it makes sense to conduct
surveys using forms. Online forms are a modern and effective way to conduct surveys. Unlike
written surveys, which are static, the questions presented in online forms can change according to
how someone responds. For instance, if you use JotForm to create your forms, when someone
answers no to a question about allergies, they won’t have to scroll past all of the related follow-up
questions about specific allergies. Instead, they’ll go immediately to a question on a different topic.
( Ainsworth, 2020)
(Source: Question Pro)

*Data audit

According to Martin (2017), data audit is considered to be a great tool for organizations to
avoid operational chaos with its main function is to improve the general quality of a data through
assessing the given data management strategy. This involves mapping key metrics then make
concise judement about the properties of a dataset.
(Source: Evolytics)

* A specific example from published sources

As for the data in this report, it has been obtained from the OPENIPCSR website
following this link. This is the result obtained from the company Axios and Ipsos, if only with the
numbers in the file of the program we will not understand what kind of data this is but just the
numbers.

According to the summary in the website from the link:


https://www.openicpsr.org/openicpsr/project/129181/version/V1/view - "Summary: Since the
COVID-19 pandemic began, Ipsos has partnered with Axios to release a weekly Coronavirus
Index survey. The survey examines Americans' behaviors, perceptions of risk, and attitudes
around COVID-19” we understand that these are survey information about human behavior of
American people. These results are collected using various survey questions about people's behavior
during the COVID pandemic, which are encoded into numbers that can be run on the SPSS system.
From there, the company Axios and Ipsos will get the amount of information and make objective
judgments when looking at the numbers in the Descriptive or Frequencies versions of the system,
which is Knowledge. In this report, the report will show 151 people in the income system with 6
different factors to assess the overall behavior of the people. From the above factors, when obtaining
the results, these two companies will cooperate with the government to come up with
timely measures and policies to support the people in the COVID-19 (Wisdom) pandemic.

III. Evaluate data from a variety of sources using different methods of analysis.

Data analysis is an important process that based firstly on raw data, through correct
methods and techniques, “mining” for necessary insights, organizations may harvest useful
information. In general, there are three distinctive methods of analysis : Descriptive, exploratory and
confirmatory.

*Descriptive analysis

According to Kenton (2019), descriptive statistics are brief descriptive coefficients that
summarize a given data set, which can be either a representation of the entire or a sample of a
population. Descriptive statistics are broken down into measures of central tendency and measures
of variability (spread). Measures of central tendency include the mean, median and mode, while
measures of variability include standard deviation, variance, minimum and maximum variables, and
kurtosis and skewness.

The most recognized types of descriptive statistics are measures of center: the mean,
median and mode, which are used at almost all levels of math and statistics. (Kenton, 2019)
+ Mean:

A mean is the simple mathematical average of a set of two or more numbers. The mean for
a given set of numbers can be computed in more than one way, including the arithmetic mean
method, which uses the sum of the numbers in the series, and the geometric mean method, which is
the average of a set of products. However, all of the primary methods of computing a simple average
produce the same approximate result most of the time. (Hayes, 2020)

Formula:

Example:

+ Median:

It is stated by Ganti (2020) that the median is the middle number in a sorted, ascending or
descending, list of numbers and can be more descriptive of that data set than the average.
Formula:

Example:

+ Mode:

The mode is the value that appears most frequently in a data set. A set of data may have one
mode, more than one mode, or no mode at all. Other popular measures of central tendency include
the mean, or the average of a set, and the median, the middle value in a set. (Hayes, 2020)
Formula:

Example:

+ How to use Descriptive analysis:

All descriptive statistics are either measures of central tendency or measures of variability,
also known as measures of dispersion. Measures of central tendency focus on the average or middle
values of data sets, whereas measures of variability focus on the dispersion of data. These two
measures use graphs, tables and general discussions to help people understand the meaning of the
analyzed data. (Kenton, 2019)
Measures of central tendency describe the center position of a distribution for a data set. A
person analyzes the frequency of each data point in the distribution and describes it using the mean,
median or mode, which measures the most common patterns of the analyzed data set. (Kenton, 2019)

Measures of variability, or the measures of spread, aid in analyzing how spread out the
distribution is for a set of data. For example, while the measures of central tendency may give a
person the average of a data set, it does not describe how the data is distributed within the set. So
while the average of the data may be 65 out of 100, there can still be data points at both 1 and 100.
Measures of variability help communicate this by describing the shape and spread of the data set.
Range, quartiles, absolute deviation and variance are all examples of measures of variability.
Consider the following data set: 5, 19, 24, 62, 91, 100. The range of that data set is 95, which is
calculated by subtracting the lowest number (5) in the data set from the highest (100). (Kenton, 2019)

*Exploratory analysis

Exploratory (versus confirmatory analysis) is the method used to explore the big data set
that will yield conclusions or predictions. According to the business analytics company Sisense,
exploratory analysis is often referred to as a philosophy, and there are many ways to approach it.
The process entails “figuring out what to make of the data, establishing the questions you want to
ask and how you’re going to frame them, and coming up with the best way to present and
manipulate the data you have to draw out those important insights.” Using exploratory analysis, data
analysts are looking for clues and trends that will help them come to a conclusion. (Notre Dame of
Maryland University, 2020)

The processes of EDA involve a myriad of tasks, including spotting mistakes and missing
data; identifying important variables in the data set; testing a hypothesis related to a specific model;
and establishing a model that can explain the data in the most succinct way possible. It also involves
determining the best way to present the final assessment. (Notre Dame of Maryland University,
2020)
Example:

+ How to use the Exploratory analysis:

The main purpose of EDA is to help look at data before making any assumptions. It can
help identify obvious errors, as well as better understand patterns within the data, detect outliers or
anomalous events, find interesting relations among the variables. Data scientists can use exploratory
analysis to ensure the results they produce are valid and applicable to any desired business outcomes
and goals. EDA also helps stakeholders by confirming they are asking the right questions. EDA can
help answer questions about standard deviations, categorical variables, and confidence intervals.
Once EDA is complete and insights are drawn, its features can then be used for more sophisticated
data analysis or modeling, including machine learning. (IBM Cloud Education, 2020)
*Confirmatory analysis

According to Statistic Solution (2020), confirmatory factor analysis (CFA) is a multivariate


statistical procedure that is used to test how well the measured variables represent the number of
constructs. Confirmatory factor analysis (CFA) and exploratory factor analysis (EFA) are similar
techniques, but in exploratory factor analysis (EFA), data is simply explored and provides
information about the numbers of factors required to represent the data. In exploratory factor
analysis, all measured variables are related to every latent variable. But in confirmatory factor
analysis (CFA), researchers can specify the number of factors required in the data and which
measured variable is related to which latent variable. Confirmatory factor analysis (CFA) is a tool
that is used to confirm or reject the measurement theory.

Example and Formula:

One Factor Confirmatory Factor Analysis

The most fundamental model in CFA is the one factor model, which will assume
that the covariance (or correlation) among items is due to a single common factor. Much like
exploratory common factor analysis, we will assume that total variance can be partitioned into
common and unique variance.

In Mplus the code is relatively simple, note the BY statement indicates the items to
the right of the statement loading onto the factor to the left of the statement.

TITLE: One Factor CFA SAQ-7 (Marker Method)

DATA: FILE IS saq8.csv;

VARIABLE: NAMES ARE q01-q08;

USEVARIABLES q01 q03-q08;

ANALYSIS: ESTIMATOR = ML;

MODEL: f1 BY q01 q03-q08;

OUTPUT: STDYX;
(Source: UCLA University)

Model Fit Statistics

The three main model fit indices in CFA are:

1. Model chi-square this is the chi-square statistic we obtain from the maximum
likelihood statistic (similar to the EFA)

2. CFI is the confirmatory factor index – values can range between 0 and 1
(values greater than 0.90, conservatively 0.95 indicate good fit)

3. RMSEA is the root mean square error of approximation (values of 0.01,


0.05 and 0.08 indicate excellent, good and mediocre fit respectively, some go up to 0.10 for
mediocre). In Mplus, you also obtain a p-value of close fit, that the RMSEA < 0.05. If you reject the
model, it means your model is not a close fitting model.
Mplus lists another fit statistic along with the CFI called the TLI Tucker Lewis
Index which also ranges between 0 and 1 with values greater than 0.90 indicating good fit. If the
CFI and TLI are less than one, the CFI is always greater than the TLI.

In our one factor solution, we see that the chi-square is rejected. This usually
happens for large samples (in this case we have N=2571). The RMSEA is 0.100 which indicates
mediocre fit. The CFI is 0.906 and the TLI is 0.859, almost but not quite at the threshold of 0.95 and
0.90.

(Source: UCLA University)


+ How to use Confirmatory analysis

CDA is the process used to evaluate evidence by challenging their assumptions


about the data. This part of the process is where they work backward from their conclusions and
weigh the merits of the results of their work. It’s like examining evidence and questioning witnesses
in a trial, trying to determine the guilt or innocence of the defendant. (Notre Dame of Maryland
University, 2020)

CDA involves processes like testing hypotheses, producing estimates, regression


analysis (estimating the relationship between variables) and variance analysis (evaluating the
difference between the planned and actual outcome). (Notre Dame of Maryland University, 2020)

*Demonstrating the pros and cons of each method about your above interpretation.

Strengths Weaknesses

Descriptive - Holistic ( Optimize both _Inaccuracy ( Respondents


quantitative and qualitative might give misleading or
research) untruthful answers )
- Practical ( Rich in information
and conductive methods, thus _Lack of reusability ( due to
there are more insights to their observational nature )
utilize)
- Accessibility to unorthodox
topic data

Exploratory - Broader and effective - Relies heavily on qualitative


predictions for the causes of
data. Hence, the result may
underlying problems
become bias
Confirmatory - Resolve common problems - Unable to test structural
equation models

To summarize, Exploratory data analysis (EDA) is the first part of data analysis process.
There are several important things to do at this stage, but it boils down to this: figuring out what to
make of the data, establishing the questions and how the analyst is going to frame them, and coming
up with the best way to present and manipulate the data which have to be drew out those important
insights. Confirmatory Data Analysis is the part where analyst evaluate evidence using traditional
statistical tools such as significance, inference, and confidence. Confirmatory Data Analysis
involves things like: testing hypotheses, producing estimates with a specified level of precision,
regression analysis, and variance analysis. In this way, the confirmatory data analysis is where the
analyst put the findings and arguments to trial. Descriptive analysis basically deals with whats
inside the data. It deals with analyzing our data-sets and driving insights. It does not deal with
coming with the right recommendation/active to solve a particular problem. Normally it involves
following steps: Bi-variate/Multi-variate analysis- Deals with simple statistical procedures such as
mean, median, standard deviation, correlation, range, variance etc. to understand data variables.
Some of the key things from this analysis is - Outliers identification, relationships between variables,
variables summary and insights. Data visualization - Data plots between different variables to drive
insights.
IV. The analysis of the research of the behaviour of people who are dealing with COVID-19 in
America

Statistics

Age Gender Have you How have the How have the How have the
personally been following following following
tested for the changed in the changed in the changed in the
coronavirus? last week, if at last week, if at last week, if at
all? - Your all? - Your all? - Your mental
emotional well physical health health
being

Valid 151 151 151 151 151 151


N
Missing 0 0 0 0 0 0

Mean 60.54 1.59 1.95 3.15 3.05 3.09

Mode 76 2 2 3 3 3

Std. Deviation 14.795 .494 .211 .651 .625 .760

The first one is a Statistic showing a summary of data taken from 151 responses to 6
different question types. We can see that the St. Deviation of the variables is not very different,
however, the difference in age variable is quite large compared to other variables up to 14,795.
Age

Frequency Percent Valid Percent Cumulative


Percent

24 1 .7 .7 .7

25 1 .7 .7 1.3

26 1 .7 .7 2.0

28 1 .7 .7 2.6

29 1 .7 .7 3.3

30 1 .7 .7 4.0

31 2 1.3 1.3 5.3

32 1 .7 .7 6.0

33 1 .7 .7 6.6

34 1 .7 .7 7.3
Valid
35 1 .7 .7 7.9

36 1 .7 .7 8.6

39 3 2.0 2.0 10.6

40 1 .7 .7 11.3

42 4 2.6 2.6 13.9

43 2 1.3 1.3 15.2

44 1 .7 .7 15.9

45 1 .7 .7 16.6

46 1 .7 .7 17.2

47 3 2.0 2.0 19.2


48 1 .7 .7 19.9

49 4 2.6 2.6 22.5

50 1 .7 .7 23.2

51 4 2.6 2.6 25.8

52 4 2.6 2.6 28.5

54 2 1.3 1.3 29.8

55 3 2.0 2.0 31.8

56 3 2.0 2.0 33.8

57 5 3.3 3.3 37.1

58 2 1.3 1.3 38.4

59 4 2.6 2.6 41.1

60 6 4.0 4.0 45.0

61 4 2.6 2.6 47.7

62 3 2.0 2.0 49.7

63 4 2.6 2.6 52.3

64 5 3.3 3.3 55.6

65 5 3.3 3.3 58.9

66 4 2.6 2.6 61.6

67 3 2.0 2.0 63.6

68 3 2.0 2.0 65.6

69 6 4.0 4.0 69.5

70 3 2.0 2.0 71.5

71 4 2.6 2.6 74.2


72 7 4.6 4.6 78.8

73 5 3.3 3.3 82.1

74 3 2.0 2.0 84.1

75 4 2.6 2.6 86.8

76 8 5.3 5.3 92.1

78 1 .7 .7 92.7

79 1 .7 .7 93.4

80 1 .7 .7 94.0

81 2 1.3 1.3 95.4

82 1 .7 .7 96.0

85 2 1.3 1.3 97.4

88 1 .7 .7 98.0

89 1 .7 .7 98.7

93 2 1.3 1.3 100.0

Total 151 100.0 100.0

The following table shows the age of the participants with the values from 24 to 93. In
which, the largest percentage is 76 (5.3%), followed by the age of 72 (4.6%). And the largest
concentration of participants is from the age of 49 to 76. This shows that the elderly in the US are
fully interested and aware of the COVID epidemic happening.
Gender

Frequency Percent Valid Percent Cumulative


Percent

Male 62 41.1 41.1 41.1

Valid Female 89 58.9 58.9 100.0

Total 151 100.0 100.0

Next is the map showing the sexes participating in the survey with only 2 genders, namely
Male and Female. Of which, 89 out of 151 are female and 62 are male. It shows that at 151 it was
found that women are more likely to participate in epidemic surveys than men.

Have you personally been tested for the coronavirus?

Frequency Percent Valid Percent Cumulative


Percent

Yes 7 4.6 4.6 4.6

Valid No 144 95.4 95.4 100.0

Total 151 100.0 100.0

In the next table you will ask if the survey participant has ever tested for COVID-19.
Surprisingly, up to 95.4% of respondents said that they had not been tested and only 4.6% said that
they had been tested. This leads to the conclusion that testing for COVID-19 in the US is currently
difficult, most probably because there is no evidence or there is no local testing site for COVID
epidemic.
How have the following changed in the last week, if at all? - Your emotional well being

Frequency Percent Valid Percent Cumulative


Percent

Improved a lot 1 .7 .7 .7

Improved a little 13 8.6 8.6 9.3

No different 105 69.5 69.5 78.8


Valid
Gotten a little worse 26 17.2 17.2 96.0

Gotten a lot worse 6 4.0 4.0 100.0

Total 151 100.0 100.0

The next table will ask about the respondents' feelings compared to the previous week. The
majority of 69.5% of respondents think that there is no change at all. 17.2% of the respondents said
that their feelings were slightly worse, 4% of the respondents said that their feelings were much
worse. In addition, 7% said that their feelings had improved a lot from the previous week, and 8.6%
said that they had slightly improved emotionally.

How have the following changed in the last week, if at all? - Your physical health

Frequency Percent Valid Percent Cumulative


Percent

Skipped 1 .7 .7 .7

Improved a lot 1 .7 .7 1.3


Valid
Improved a little 10 6.6 6.6 7.9

No different 119 78.8 78.8 86.8


Gotten a little worse 17 11.3 11.3 98.0

Gotten a lot worse 3 2.0 2.0 100.0

Total 151 100.0 100.0

In addition, the next version will ask if the respondents' physical condition has changed
much from the previous week. 78.8% of respondents said that there is no change, 1 respondent
ignored this answer. Moreover, 17 respondents said that their physical condition was slightly worse.

How have the following changed in the last week, if at all? - Your mental health

Frequency Percent Valid Percent Cumulative


Percent

Skipped 2 1.3 1.3 1.3

Improved a lot 1 .7 .7 2.0

Improved a little 10 6.6 6.6 8.6

Valid No different 108 71.5 71.5 80.1

Gotten a little worse 26 17.2 17.2 97.4

Gotten a lot worse 4 2.6 2.6 100.0

Total 151 100.0 100.0

Finally, the question about psychology of the respondents, surprisingly, there are 17.2% of
respondents that their psychology is getting worse and only 1 of the respondents think that their
psychology is improving. 71.5% of respondents said that there is no change in their psychology in
the past weeks.
To conclude, although this is only a survey of 151 people, it has shown somewhat about
the impact of the COVID-19 epidemic on the American people, both physically and mentally, in
recent times. If the results are based on a broader range, more accurate numbers and data are
possible.

V. Conclusion and suggestion

In conclusion, it can be said that informational data will help us gain knowledge, from that
knowledge along with the judgments can suggest measures as well as methods to minimize risks as
well as avoid possible variables. As a research analyst, in my opinion, media companies should
provide public surveys and information to the government in order to coordinate propaganda against
contagion. COVID-19 in the community. In addition, the government can base on the above data to
provide options as well as guide people to test to control the level of spread in the community.
Reference list:

Openicpsr.org. 2020. Axios-Ipsos Coronavirus Index. [online] Available at:


<https://www.openicpsr.org/openicpsr/project/129181/version/V1/view> [Accessed 2 February 2021].

Guru99.com. 2020. Difference between Information and Data. [online] Available at:
<https://www.guru99.com/difference-information-data.html> [Accessed 2 February 2021].

Byjus.com. 2020. Difference Between Data and Information. [online] Available at:
<https://byjus.com/biology/difference-between-data-and-information/> [Accessed 2 February 2021].

Ag Data Commons U.S. DEPARTMENT OF AGRICULTURE. 2020. Data Dictionary - Examples. [online]
Available at: <https://data.nal.usda.gov/data-dictionary-examples> [Accessed 2 February 2021].

Brahmachary, A., 2020. DIKW Model: Explaining the Concept of DIKW Hierarchy in ITIL. [online]
CertGuidance. Available at: <https://www.certguidance.com/explaining-dikw-hierarchy/> [Accessed 2
February 2021].

Spiderum. 2020. Mô hình tháp DIKW - Con đường tới Sự Thông Thái ^^!. [online] Available at:
<https://spiderum.com/bai-dang/Mo-hinh-thap-DIKW-Con-duong-toi-Su-Thong-Thai-pe2> [Accessed 2
February 2021].

Ainsworth, Q., 2020. Data Collection Methods. [online] JotForm. Available at:
<https://www.jotform.com/data-collection-methods/> [Accessed 2 February 2021].

Evolytics. 2020. A Data Audit will Reveal Analytics Gaps. [online] Available at:
<https://evolytics.com/blog/data-audit-reveal-analytics-gaps/> [Accessed 2 February 2021].

Kenton, W., 2019. Descriptive Statistics. [online] Investopedia. Available at:


<https://www.investopedia.com/terms/d/descriptive_statistics.asp> [Accessed 2 February 2021].

Hayes, A., 2020. Mean. [online] Investopedia. Available at:


<https://www.investopedia.com/terms/m/mean.asp> [Accessed 2 February 2021].

Ganti, A., 2020. Median Definition. [online] Investopedia. Available at:


<https://www.investopedia.com/terms/m/median.asp> [Accessed 2 February 2021].

Hayes, A., 2020. Mode Definition. [online] Investopedia. Available at:


<https://www.investopedia.com/terms/m/mode.asp> [Accessed 2 February 2021].

Ibm.com. 2020. What is Exploratory Data Analysis?. [online] Available at:


<https://www.ibm.com/cloud/learn/exploratory-data-analysis> [Accessed 2 February 2021].
Statistics Solutions. 2020. Confirmatory Factor Analysis - Statistics Solutions. [online] Available
at: <https://www.statisticssolutions.com/confirmatory-factor-analysis/> [Accessed 2 February 2021].

NDMU Online. 2020. Exploratory Analysis vs. Confirmatory Analysis. [online] Available at:
<https://online.ndm.edu/news/analytics/exploratory-analysis-vs-confirmatory-analysis/> [Accessed 2
February 2021].

Stats.idre.ucla.edu. 2020. A Practical Introduction to Factor Analysis: Confirmatory Factor Analysis.


[online] Available at: <https://stats.idre.ucla.edu/spss/seminars/introduction-to-factor-analysis/a-
practical-introduction-to-factor-analysis-confirmatory-factor-analysis/> [Accessed 2 February 2021].

You might also like