The Four Main Aspects of Statistics Can Be Further Classified Into Subcategories As Follows

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

The four main aspects of statistics can

be further classified into subcategories


as follows:

1. Descriptive statistics:
Frequency distributions
Central tendency measures (mean, median, mode)
Dispersion measures (range, variance, standard deviation)
Graphical representation (histograms, bar graphs, scatter plots)
2. Probability theory:
Sample space and event space
Axioms of probability
Conditional probability
Bayes' theorem
3. Inferential statistics:
Estimation of population parameters
Hypothesis testing
Confidence intervals
T-tests, ANOVA, chi-square tests, regression analysis
4. Statistical modeling:
Linear regression
Logistic regression
Decision trees
Naive Bayes
Random forest
Neural networks

Note: This is just a brief overview and the subcategories can further be divided into
more specific areas of study.

1. Descriptive statistics:
Frequency distributions: The number of times each value in a dataset
appears. For example, the number of times each score appears on a test.
Central tendency measures: Statistics that describe where the center of the
data lies, such as mean, median, and mode. For example, the average
height of a group of people.
Dispersion measures: Statistics that describe how spread out the data is,
such as range, variance, and standard deviation. For example, the spread
of salaries within a company.
Graphical representation: Visual representation of data to understand
patterns, trends, and relationships, such as histograms, bar graphs, and
scatter plots. For example, a histogram of exam scores to show the
distribution.
2. Probability theory:
Sample space and event space: The set of all possible outcomes of an
experiment and a subset of those outcomes, respectively. For example, the
sample space could be the number of possible outcomes of rolling a die,
and an event could be rolling a specific number.
Axioms of probability: The basic rules that govern the calculation of
probability. For example, the probability of an event must be between 0
and 1.
Conditional probability: The probability of an event given that another
event has occurred. For example, the probability of getting a heads on a
coin flip given that it's tails up.
Bayes' theorem: A method for updating the probability of an event based
on new information. For example, updating the probability of a medical
condition given a positive test result.
3. Inferential statistics:
Estimation of population parameters: The process of using sample data to
estimate the values of parameters in a population. For example, estimating
the average weight of all fish in a lake from a sample of fish.
Hypothesis testing: The process of using sample data to test a claim about
a population. For example, testing the claim that a new drug is effective at
reducing blood pressure.
Confidence intervals: A range of values that is likely to contain the true
value of a population parameter with a certain degree of confidence. For
example, finding the range of values that is likely to contain the true
average weight of fish in a lake with 95% confidence.
T-tests, ANOVA, chi-square tests, regression analysis: Methods for
comparing means, variances, proportions, and relationships between
variables, respectively. For example, using a t-test to compare the means
of two groups, using ANOVA to compare the means of three or more
groups, using a chi-square test to compare proportions, and using
regression analysis to model the relationship between two variables.
4. Statistical modeling:
Linear regression: A method for modeling the relationship between a
dependent variable and one or more independent variables. For example,
modeling the relationship between a person's height and weight.
Logistic regression: A method for modeling the relationship between a
dependent binary variable and one or more independent variables. For
example, modeling the relationship between a person's age and likelihood
of getting a disease.
Decision trees: A method for modeling decisions or predictions based on
certain conditions. For example, modeling the decisions made by a doctor
based on a patient's symptoms.
Naive Bayes: A probabilistic method for classifying data into different
categories. For example, classifying emails as spam or not spam.
Random forest: An ensemble method for combining multiple decision
trees to make a prediction. For example, making a prediction of whether a
person will default on a loan based on their credit history and other factors.
Neural networks: A type of machine learning model that is inspired by the
structure and function of the human brain. For example, recognizing
handwritten digits based on images of handwritten digits.

You might also like