Professional Documents
Culture Documents
Design of Experiment Project Report
Design of Experiment Project Report
Design of Experiment Project Report
1.1 ABSTRACT
There are different approaches to the problem of assigning each word of a text with a
parts-of-speech tag, which is known as Part-Of-Speech (POS) tagging. In this paper
we compare the performance of a few POS tagging techniques e.g. statistical approach
(n-gram, HMM) and transformation based approach (Brill’s tagger) and Baum Welch
Approach. A supervised POS tagging approach requires a large amount of annotated
training corpus to tag properly. At this initial stage of POS-tagging. We tried to see which
technique maximizes the performance with this limited resource.
1.2 INTRODUCTION
1.2.1 Part-Of-Speech Tagging
POS tagging refers labelling the word corresponding to which POS best describes the
use of the word in the given sentence. Part-Of-Speech refers to the purpose of a
word in a given sentence.
However, the POS tag of a word can vary depending on the context in which it is
used. One way out of this is to make use of the context of occurrence of a word. As
mentioned above, the POS tag depends on the context of its use. There are set of
rules for some POS tags dictating what POS tag should follow or precede them in a
sentence. For example, a word that occurs between an determiner and a noun
should be an adjective.
1
PROJECT REPORT
1.3 METHODOLOGY
1.3.1 Hidden Markov Model
Hidden Markov Model is a probabilistic sequence model, that computes
probabilities of sequences based on a prior and selects the best possible sequence
that has the maximum probability. Sometimes, what we want to predict is a
sequence of states that aren’t directly observable in the environment. Though we are
given another sequence of states that are observable in the environment and these
hidden states have some dependence on the observable states. If you notice closely,
we can have the words in a sentence as Observable States (given to us in the data)
but their POS Tags as Hidden states and hence we use HMM for estimating POS tags.
It must be noted that we call Observable states as ‘Observation’ & Hidden states as
‘States’.
2
PROJECT REPORT
Our objective is to find the sequence that maximizes the probability defined in the above
diagram.
Rule Based Taggers Disambiguation is done by analyzing the linguistic features of the word, its
preceding word, its following word, and other aspects.
Stochastic Taggers The tag encountered most frequently in the training set with the word is the
one assigned to an ambiguous instance of that word.
Transformation Based allows us to have linguistic knowledge in a readable form, transforms
one state to another state by using transformation rules. It draws the inspiration from both the
previous explained taggers − rule-based and stochastic.
3
PROJECT REPORT
4
PROJECT REPORT
5
PROJECT REPORT
Example:
POS tags assigned to the words in the article using Viterbi algorithm.
6
PROJECT REPORT
1.3.9 ANALYSIS
In our experimental design, we have 3 algorithms as treatments, and 4 newspapers
as blocks to make a RANDOMIZED BLOCK DESIGN (RBD).
HYPOTHESIS:
o Null hypothesis: H0: All treatments are homogeneous Ta=Tb=Tc=Td=Te =Tf
o Null hypothesis: Ho1: Pair of treatments have the same effect i.e., αi = αi’
o Alternate hypothesis: H11: Pair of treatment have different effects i.e., αi ≠ αi’
7
PROJECT REPORT
ANOVA TABLE
Algorithms
61.268 2 30.634 174.613 .000
(Treatments)
Total 111621.740 12
Since p-value for treatments is less than 0.05, we have sufficient evidence to reject the null
hypothesis Ho and can conclude that the treatments are different.
8
PROJECT REPORT
INDIVIDUAL DIFFERENCES
Multiple Comparisons
Brill- 6.4211
5.5123* .29617 .000 4.6036
tagger
Viterbi
Baum- 4.0964
3.1877* .29617 .000 2.2789
Welch
- -
Viterbi -5.5123* .29617 .000 4.6036
Brill- 6.4211
tagger Baum- - -
-2.3247* .29617 .001 1.4159
Welch 3.2334
- -
Viterbi -3.1877* .29617 .000 2.2789
Baum- 4.0964
Welch Brill-
2.3247* .29617 .001 1.4159 3.2334
tagger
Based on observed means.
The error term is Mean Square(Error) = .175.
On comparing individual differences, we can see that the mean difference is significant
for all pairs as p-value < 0.05, and we may reject the null hypothesis Ho1, and conclude
that all pairs are significantly different.
9
PROJECT REPORT
1.4 CONCLUSION
The RBD experimental design yielded that the accuracy of Parts-of-Speech tagging
algorithms is significantly different for different newspapers. Also, the individual pair
differences are also significantly different for all the algorithms.
1.5 BIBLIOGRAPHY
o https://medium.com/@zhe.feng0018/coding-viterbi-algorithm-for-hmm-from-scratch-
ca59c9203964
o https://www.nltk.org/
o https://www.projectpro.io/recipes/what-is-brill-tagger
o http://www.adeveloperdiary.com/data-science/machine-learning/derivation-and-
implementation-of-baum-welch-algorithm-for-hidden-markov-model/
10