Professional Documents
Culture Documents
Sentiment Analysis For PolishPoznan Studies in Contemporary Linguistics
Sentiment Analysis For PolishPoznan Studies in Contemporary Linguistics
445–468
© Faculty of English, Adam Mickiewicz University, Poznań, Poland
doi: 10.1515/psicl-2019-0016
ALEKSANDER WAWER
Institute of Computer Science, Polish Academy of Sciences, Warsaw
axw@ipipan.waw.pl ORCID 0000-0002-7081-9797
ABSTRACT
This article is a comprehensive review of freely available tools and software for senti-
ment analysis of texts written in Polish. It covers solutions which deal with all levels of
linguistic analysis: starting from word-level, through phrase-level and up to sentence-
level sentiment analysis. Technically, the tools include dictionaries, rule-based systems
as well as deep neural networks. The text also describes a solution for finding opinion
targets. The article also contains remarks that compare the landscape of available tools in
Polish with that for English language. It is useful from the standpoint of multiple disci-
plines, not only information technology and computer science, but applied linguistics
and social sciences.
KEYWORDS: Sentiment analysis; opinion extraction; opinion targets; deep neural net-
works.
1. Introduction
The purpose of this article is to review various tools available for Polish lan-
guage sentiment analysis and related areas. The understanding of sentiment
analysis in this text is a very broad one: it consists of distinguishing whether a
document, sentence, phrase or word is positive, negative or neutral. The field is
closely related to opinion mining and opinion classification.
It is worth noting that multiple different definitions of sentiment analysis
and opinion mining are in use. In his book, Bing Liu (Liu 2012) defines an opin-
ion as a quintuple composed of an entity, it’s aspect, a sentiment on this aspect,
a holder and a time when it’s expressed. The sentiment is positive, negative, or
neutral, or expressed with different strength/intensity levels. In this understand-
446 A. Wawer
In this section we discuss the most basic building blocks of word-level senti-
ment analysis: dictionaries. Typically, these resources assign sentiment values
(for example labels: positive, neutral, negative) to lexical units, often word base
forms. In Section 2.1 we cover automatically generated dictionaries, in Section
2.2 manually created dictionaries are discussed.
1
http://zil.ipipan.waw.pl/SlownikWydzwieku
Sentiment analysis for Polish 449
Sentiment dictionaries may also be compiled from corpora with sentiment labels
attached to documents, such as customer reviews with ratings. Positive reviews
tend to contain more positively oriented words, negative words appear more of-
ten in negative reviews. A method of sentiment dictionary acquisition which ex-
plores this intuition has been described in Haniewicz et al. (2013). Unfortunate-
ly, the dictionary discussed in this article is not publicly available. It is based on
another resource (a semantic network), also not publicly available. The reported
evaluation of the dictionary is performed on reviews from the same domain as
those used to generate the dictionary and reaches 0.79 in terms of accuracy
computed on the level of reviews. Word-level accuracy is likely lower.
For languages other than Polish, other approaches have been proposed. For
example, Rill et al. (2012) introduced a method to derive sentiment phrases
from reviews. Authors generated lists of opinion bearing phrases with their
opinion values in a continuous range between −1 and 1.
The only publicly available, manually created sentiment Polish language dic-
tionary is the one described in (Zaśko-Zielińska et al. 2015). The dictionary, as
described in the article, is annotated on the level of lexical units from plWord-
Net. The dictionary contains sentiment annotations of 32 thousands lexical
units: 19.6 for nouns and 11.6 for adjectives. It has five polarity labels (strong
positive, weak positive, neutral, weak negative, strong negative).
One interesting property of this dictionary is the fact that it allows a word to
have multiple, possibly even opposing sentiment labels depending on its sense.
This is done via linking the units to their synsets, determined by word senses
and synonymy relation. In most cases this turns out to be sufficient. However,
about 5% of the lexical units were marked as ambiguous. Even within their
senses as determined by synonymy relation and synsets, annotators identified
various sentiment polarities.
The usage of this dictionary to label the text (assign word sentiments) re-
quires determining senses of ambiguous words by analyzing sentence- or even
document-level context of their appearances. In automated analysis, a word
sense disambiguation algorithm should decide which sense has been used in a
450 A. Wawer
The state of dictionaries can be summarized as follows. There are two available
resources, each with some drawbacks. The first is one automatically generated
with lexeme-level sentiment assignments. It suffers from a relative lower quality
of sentiment labellings, but is usable for automated text processing. The second
one is a dictionary with manual word sense-level sentiment assignments. Its
quality is unquestionably better, but it is less feasible for automated use due to
word sense disambiguation issues described previously.
Possibly, the best choice for real world usage is to combine both dictionaries
into one list, make sure that sentiment labels correspond to the most frequent
sense of each word. This approximates lexeme-level annotation.
2
http://ws.clarin-pl.eu/wsd.shtml
3
http://nlp.pwr.wroc.pl/redmine/projects/nlprest2/wiki
Sentiment analysis for Polish 451
4
http://multiservice.nlp.ipipan.waw.pl/
5
http://git.nlp.ipipan.waw.pl/multiservice/clients
452 A. Wawer
There is one more rule type (which isn’t mentioned here) and possible in Senti-
pejd: it is simply a listing of all multi-word expressions that carry some senti-
ment as a whole, while none of the words have any sentiment when considered
individually. This scenario is similar to idiomatic expression detection.
The application of rules in Buczyński and Wawer (2008b) demonstrated to
increase the review classification accuracy up to four percent, depending on set-
ting. The effect was more noticeable in the case of more wordy, longer product
reviews.
To sum up, Sentipejd provides an improvement over word-level dictionary
only analysis via extension into phrase-level sentiment analysis. It is a highly
versatile tool that can quite easily be adapted into many domains by changing or
fine-tuning its sentiment lexicon. Rules, as they are now implemented, are not
overly lexicalized and should be easy to re-use.
Its main drawback is no support for sentence-level sentiment and handling
more complex phrasal structures. It is not clear how to combine sentiment from
multiple sentiment sources. For example, it is not clear how to combine senti-
ment from a negative word and positive phrase which both occurred in the same
sentence. Which takes precedence?
This boils down to the fact that grammars, which empower Spejd, are shal-
low and often there is no single, usable syntactic link between such phrases and
words (which would create a rule for deciding which takes precedence). It is
hard, if not entirely impossible, to enumerate all possible cases using shallow
grammar rules such as Sentipejd’s. The difficulty becomes compounded due to
relatively free word order when constructing Polish sentences.
Sentiment analysis for Polish 453
Solutions have been proposed to handle this problem in modern day senti-
ment analysis (also in Polish) as presented in Section 4. The next section reveals
tools suitable for computing sentiment of any syntactically coherent phrase,
provided that it can be formed using a sentence’s dependency parse graph.
4.1. Introduction
This section presents tools that are suitable not only for all previously men-
tioned tasks (word-level and phrase-level sentiment analysis) but additionally
also for sentence-level analysis. As previously mentioned, these tools can pre-
dict sentiment of almost any phrase, including much more complex phrases than
Sentipejd. The only restriction on phrase type is syntactic: phrases have to be
formed from sub-trees of sentence’s dependency tree graphs.
There are several specific features that all tools presented here share.
First of all and most broadly, they are all based on deep learning. This no-
tion is used to describe multi-layer neural networks that became heavily popular
in natural language processing after 2013.
Arguably, their popularity can be explained at least partially by the seminal
paper by Mikolov et al. (2013). The study demonstrated interesting regularities
in word representations learned using neural networks. The representations,
called word embeddings, were very promising in solving word analogy tasks us-
ing simple arithmetics. In the famous example, arithmetic operation on vectors
of three words: “King-Man+Woman” results in a vector very close to “Queen”.
It indicates that words with similar meaning tend to have similar word embed-
ding vectors. It quickly became obvious that using such word representations
combined with multi-layer neural networks trained on very large corpora allow
science to match (or in some cases surpass) the existing state-of-the-art in many
areas of natural language processing.
Second, within the broad area of deep learning, all tools presented in this
section are based on recurrent neural networks (RNNs), and their sub-type such
as long short-term memory networks (LSTMs) (Hochreiter and Schmidhuber
1997). RNNs are currently fundamental building blocks not only for sentiment
analysis but for many other language processing tasks, including machine trans-
lation, language modeling and question answering.
Third, as is the case with most natural processing tools of this type, they use
pre-trained word embedding vectors to represent a word’s meaning.
454 A. Wawer
Four, the tools share the same task. Given a dependency tree of a sentence,
the goal is to provide the correct sentiment for each possible sub-tree (phrase).
Phrases correspond to sub-trees of a dependency parse tree. By convention, as
discussed in Section 4.2, sentiment values are assigned to whole phrases (and in
some cases, whole sentence), regardless of their type. Typically, applications
compute sentiment recursively, starting from leaves and smaller phrases, then
expanding to larger phrases and taking into account sentiment values already
computed for their nested sub-phrases. This could be equivalent to recursively
folding the tree in a bottom-up fashion. Sentence-level sentiment becomes then
the value of predictive model after folding the whole sentence.
The next section briefly discusses the resource used to train the neural net-
work models, and provides insight into data format provided at output.
The dataset used to train neural network models described in Section 4.3 is a
dependency treebank with sentiment annotations. It is available for downloaded
from <http://zil.ipipan.waw.pl/TreebankWydzwieku>.
The inspiration for the Polish Sentiment Treebank comes from the English
language, where multiple tools and approaches emerged after publishing the
Stanford Sentiment Treebank (Socher et al. 2013). This English reference data
set contains 9645 sentences from a movie review domain. It has been widely
used for evaluating multiple deep learning approaches, such as simple recursive
neural networks and recently more complex Tree LSTMs (Tai et al. 2015a).
For each sentence, its overall sentiment (neutral, positive and negative), as
well as sentiments of each sub-phrase (sub-tree) and each leaf word have been
assigned by a linguist.
Sentiment annotations for each token corresponded to the overall sentiment
of the whole phrase under it and included within it. Specifically:
– For every leaf token or word, its sentiment corresponds to this word or to-
ken’s sentiment.
– For every non-leaf token or word (node that has non-empty set of children)
sentiment field describes the sentiment of the whole phrase, formed by sub-
tree starting at this token (that includes this token and all tokens below it)
two parts. The first part of the treebank is composed from the sub-part of Sklad-
nica treebank, namely from sentences that contain at least one sentiment-bearing
word. This part consists of 235 sentences (1915 sentiment-annotated multiword
phrases). The second part consits of 965 sentences from a product review corpus
available from <http://zil.ipipan.waw.pl/OPTA/>. The number of sentiment-
annotated multiword phrases is 4640.
The dataset is the first Polish language corpus with fully labeled parse trees
that allows for analysis of compositional effects of sentiment in Polish. It cap-
tures complex sentiment-related phenomena at the intersection of syntax and
semantics. The dataset enables us to train compositional models that are based
on supervised and structured machine learning techniques. It is especially well-
suited for recurrent neural networks.
The dataset was used extensively during the PolEval 2017 competition as
training and test data for participating systems. It provided the opportunity to
design, train and compare such algorithms on the Polish language data sets.
Figure 1 presents one example sentence from the Polish Sentiment Tree-
bank.6 The figure illustrates also dependency parse structure. Only selected
phrase-level (sub-trees) sentiment annotations have been presented. The sen-
tence is a mixture of both positive and negative sentiments, with negative sen-
timents taking over the precedence over positive. Positive sentiments (‘a year
ago was really ok’) were overridden by more recent negative experiences (‘I do
not recommend’). Clearly, understanding the overall sentiment composition re-
quires capturing the time frame of both experiences.
The next section discusses systems submitted to the sentiment analysis subtask
of the Poleval 2017 competition.
6
The English translation of the sentence is as follows: ‘Generally I do not recommend, but a year
ago was really ok’.
456 A. Wawer
data. The selection of sentences into train and validation sets in predictions 1
and 2 was a bit different in terms of sizes and selection methods.
– predictions 3: The training phase length was determined using early stopping
condition set to 5 epochs, measured on test data. Then, we re-trained this
model on the combined train and validation sets for 15 more epochs.
Like all other models in the PolEval sentiment subtask, the model was created
using LSTM architecture of recurrent deep neural network in tree topology. For
training, the authors used syntactic dependency trees without dependency labels.
The model was built as follows:
– Leaves of NN are embeddings of leaves in dependency tree.
– Subtrees of NN are embeddings of subtrees in dependency tree and are cre-
ated by combining embedding of a root of dependency subtree one-by-one
with its children (from left to right) using LSTM equations.
– Embedding of each subtree was projected on its sentiment label.
The system used in this solution was a Child-Sum Tree-LSTM deep neural net-
work (as described by (Tai et al. 2015a)), fine-tuned for dealing with morpho-
logically rich languages. Fine-tuning included applying a custom regularization
technique (zoneout, described by Krueger et al. 2016), and further adapted for
7
https://github.com/norbertryc/poleval
8
https://github.com/michal-lew/tree-lstm
458 A. Wawer
Given the files with results of dependency syntax analysis for each sentence,
participants of PolEval 2017 were to provide sentiment labelings. They were to
provide sentiment values of specific words for leaves. For non-leaf tokens they
were to provide sentiment values that reflect the sentiment of the overall phrase,
formed by sub-tree starting at this token (that includes this token and all tokens
below it). Sentiment labelings obtained from participants were compared with
the organizer’s test set of gold (reference) annotation.
To evaluate the performance of the systems submitted to Poleval, their mi-
cro accuracy has been calculated against the provided test data. Micro accuracy
means that each sub-tree has been taken into account when calculating scores. It
is a global sum of correctly labeled sentiment scores, over cases belonging to
each class.10
The official results measured at the PolEval competition are presented in
Table 1.
The competing systems were heavily tuned to obtain the highest possible
prediction quality using such regularization techniques as dropout, zone-out or
L2. The winning system Tree-LSTM-NR was a Tree LSTM variant with infor-
mation from dependency labels and word2vec (Mikolov et al. 2013) embed-
dings generated for Polish from Wikipedia and the National Corpus of Polish
(www.nkjp.pl).
9
https://github.com/tomekkorbak/treehopper
10
Micro is opposed to macro accuracy, where for each class (positive, neutral, negative) its accura-
cy is computed separately, to later compute the average of class scores.
Sentiment analysis for Polish 459
Table 2. The top five most frequent dependency patterns: precision (P),
number of correct (C) and erroneous (E) matches.
Path P C E
[pos:adj] <adjunct [pos:subst] 0.886 396 51
[pos:fin] >comp [pos:prep] >comp [pos:subst] 0.814 48 11
[pos:adj] >adjunct [pos:prep] >comp [pos:subst] 0.906 48 5
[pos:adj] <adjunct [pos:subst] >adjunct [pos:prep] >comp [pos:subst] 0.333 16 32
[pos:adj] <pd [pos:fin] >subj [pos:subst] 0.909 40 4
matches for a token (ruleAny) or by indicating the particular rule that matched
(ruleID). Experiments have also been carried out to investigate the influence of
adding lexical information (lemma).
Table 3 presents the results of various input feature spaces for the CRF pre-
dictions of target words. The results are average values of precision (P), recall
(R) and F1 measured in 10-fold cross validation. A more comprehensive version
of the table can be found in Wawer (2015).
Features P R F1
lemma 0.586 0.33 0.421
lemma + POS + dep 0.553 0.466 0.505
POS + dep 0.548 0.426 0.478
POS + dep + ruleAny 0.783 0.891 0.833
POS + dep + ruleAny + ruleID + S 0.829 0.889 0.857
The main conclusion is that models without lexical features (lemma) outper-
formed ones with lexical features by large margin. The value of this result is that
it carries a hope of domain independence: machine learning from syntactic
structures appears to be much more universally applicable than lexicalized vari-
ants, attached to specific words that are usually tied to some domain. Generally
the reported CRF results (especially the models that use rule-based features)
leave little room for further improvements due to their high overall perfor-
mance.
The best performing feature set, namely the last line in Table 3 has been im-
plemented in the Python language and made available publicly as a tool called
OPFI. The package requires that the input was dependency parsed and pre-
annotated with sentiment on word-level. The first requirement can be fulfilled
using the dependency parser Malt and models that are available for download
from <http://zil.ipipan.waw.pl/PDB/PDBparser>. The second word-level senti-
ment annotations are possible using one of the dictionaries described in Section
2. Using rule-based sentiments is also possible, but not advisable: sentiment po-
larities are not relevant for opinion targets. What matters is the placement of
sentiment-bearing words, regardless of their phrase-level or contextual polarity
modifications.
Sentiment analysis for Polish 463
This section contains a short comparison with tools available in English. The
language selected here as a reference is without doubts, the one with most avail-
able tools and resources. By comparing with the English language, one can
clearly see what is missing for Polish. The review in this section is by no means
comprehensive and focuses only on major areas, as more detailed comparison
would fall out of the scope of this article.
The first area, entirely missing in Polish and closely related to sentiment
analysis, is subjectivity detection. In English, the development of subjectivity-
related field has been facilitated by corpora such as MPQA (Deng and Wiebe
2015) and IMDB subjectivity dataset (Pang and Lee 2004). One of these tools is
OpinionFinder (Wilson et al. 2005). It offers not only sentiment analysis in the
sense of polarity recognition, but also subjectivity analysis: automatically iden-
tifying when opinions, sentiments, speculations, and other private states are pre-
sent in text. More recent applications offering sentence-level subjectivity classi-
fication are usually based on bidirectional recurrent neural networks.12
Another area not addressed in Polish is document-level sentiment classifica-
tion and more generally, the problem of computing sentiment of longer utter-
ances. In English, available tools are usually based on machine learning para-
digm, models often trained on the IMDB movie review corpus (Pang and Lee
2004; Pang et al. 2002). This data set serves as an example in multiple machine
learning and sentiment analysis tutorials that demonstrate how to build a docu-
ment-level sentiment classification tools, including the well-known deep learn-
ing package Keras.13
11
http://zil.ipipan.waw.pl/OPTA
12
For example https://github.com/fractalego/subjectivity_classifier
13
The source code of the example is available for download from:
https://github.com/keras-team/keras/blob/master/examples/imdb_lstm.py
464 A. Wawer
7. Summary
The article presents a review of publicly available tools and selected resources
for sentiment analysis in Polish, covering word-level, phrase-level and sentence-
level analysis. Tools and resources known from literature but not publicly avail-
able were mentioned only briefly or entirely skipped. The article reflects the
state of knowledge of its authors as of mid 2018. It is entirely possible that we
were not aware of some items that fulfill our criteria and should be described
here or were under preparation at the time of writing, and therefore not yet in-
dexed by search engines and scholarly databases.
The list of applications and resources described more thoroughly includes:
two publicly available dictionaries, one rule-based phrase-level tool, three deep
neural network solutions for phrase-level and sentence-level analysis. Finally, it
also contains a description of an application for matching sentiments with their
targets.
The main focus was to present available software applications. Only senti-
ment dictionaries were taken into account in a manner similar to applications
because of their usability to compute sentiment with very little programming ef-
fort. In order to develop a full-fledged dictionary-based sentiment analysis solu-
tion in Polish, one would need to supplement it with a lemmatizer (morphosyn-
tactic disambiguator) such as Pantera (Acedański 2010) integrated in Multi-
Sentiment analysis for Polish 465
14
http://multiservice.nlp.ipipan.waw.pl/en/
15
http://ws.clarin-pl.eu/
466 A. Wawer
REFERENCES
Acedański, S. 2010. “A morphosyntactic brill tagger for inflectional languages”. In:
Loftsson, H., E. Rögnvaldsson and S. Helgadóttir (eds.), Advances in natural lan-
guage processing. Berlin, Heidelberg: Springer. 3–14.
Bojanowski, P., E. Grave, A. Joulin and T. Mikolov. 2016. “Enriching word vectors
with subword information”. arXiv preprint arXiv:1607.04606.
Buczyński, A. and A. Wawer. 2008a. “Shallow parsing in sentiment analysis of product
reviews”. Proceedings of the Partial Parsing Workshop at LREC. 14–18.
Buczyński, A. and A. Wawer. 2008b. “Automated classification of product review sen-
timents in Polish”. Proceedings of the Intelligent Information Systems (IIS).
Deng, L. and J. Wiebe. 2015. “MPQA 3.0: An entity/event-level sentiment corpus”. In:
Mihalcea, R., J. Yue Chai and A. Sarkar (eds.), HLT-NAACL. The Association for
Computational Linguistics. 1323–1328.
Haniewicz, K., W. Rutkowski, M. Adamczyk and M. Kaczmarek. 2013. “Towards the
lexicon-based sentiment analysis of Polish texts: Polarity lexicon”. In: Bǎdicǎ, C.,
N. Thanh Nguyen and M. Brezovan (eds.), Computational collective intelligence.
Technologies and applications. Berlin: Springer. 286–295.
Hochreiter, S. and J. Schmidhuber. 1997. “Long short-term memory”. Neural Computa-
tion 9(8). 1735–1780. doi:10.1162/neco.1997.9.8.1735.
Kędzia, P., M. Piasecki and M. Orlińska. 2015. “Word sense disambiguation based on
large scale polish clarin heterogeneous lexical resources”. Cognitive Studies 15.
269–292.
Kloumann, I. M., C. M. Danforth, K. D. Harris, C. A. Bliss and P. S. Dodds. 2012. “Pos-
itivity of the English language”. PloS One 7(1). e29484.
Krueger, D., T. Maharaj, J. Kramár, M. Pezeshki, N. Ballas, N. R. Ke, A. Goyal, et al.
2016. “Zoneout: Regularizing rnns by randomly preserving hidden activations”.
CoRR abs/1606.01305. <http://arxiv.org/abs/1606.01305>
Lafferty, J. D., A. McCallum and F. C. N. Pereira. 2001. “Conditional random fields:
Probabilistic models for segmenting and labeling sequence data”. Proceedings of
the Eighteenth International Conference on Machine Learning (ICML ’01). San
Francisco, CA: Morgan Kaufmann Publishers Inc. 282–289.
Levy, R. and G. Andrew. 2006. “Tregex and tsurgeon: Tools for querying and manipu-
lating tree data structures”. Proceedings of the Fifth International Conference on
Language Resources and Evaluation. 2231–2234.
Lew, M. and P. Pęzik. 2017. “A sequential child-combination tree-lstm network for sen-
timent analysis”. In Vetulani, Z. (ed.), Proceedings of the 8th Language and Tech-
nology Conference: Human Language Technologies as a Challenge for Computer
Science and Linguistics. Poznań.
Liu, B. 2012. Sentiment Analysis and Opinion Mining. Morgan and Claypool Publishers.
Mikolov, T., K. Chen, G. Corrado and J. Dean. 2013. “Efficient estimation of word rep-
resentations in vector space”. CoRR abs/1301.3781.
<http://arxiv.org/abs/1301.3781>.
Pang, B. and L. Lee. 2004. “A sentimental education: Sentiment analysis using subjec-
tivity summarization based on minimum cuts”. Proceedings of ACL. 271–278.
Sentiment analysis for Polish 467
Pang, B., L. Lee and S. Vaithyanathan. 2002. “Thumbs up? Sentiment classification us-
ing machine learning techniques”. CoRR cs.CL/0205070.
<http://arxiv.org/abs/cs.CL/0205070>
Przepiórkowski, A. and A. Buczyński. 2007. “Spade: Shallow Parsing and Disambigua-
tion Engine”. Proceedings of the 3rd Language and Technology Conference
(LTC’07). Poznań.
Radziszewski, A. 2013. “A tiered crf tagger for Polish”. In: Bembenik, R., Ł. Skoniecz-
ny, H. Rybiński, M. Kryszkiewicz and M. Niezgódka (eds.), Intelligent tools for
building a scientific information platform: Advanced architectures and solutions.
Berlin: Springer. 215–230.
Rill, S., J. Scheidt, J. Drescher, O. Schütz, D. Reinel and F. Wogenstein. 2012. “A ge-
neric approach to generate opinion lists of phrases for opinion mining applications”.
Proceedings of the first international workshop on issues of sentiment discovery
and opinion mining (WISDOM ’12). New York: ACM. 7:1–7:8.
Ryciak, N. 2017. “Polish language sentiment analysis with tree-structured long short-
term memory network”. In: Vetulani, Z. (ed.), Proceedings of the 8th Language and
Technology Conference: Human Language Technologies as a Challenge for Com-
puter Science and Linguistics. Poznań.
Socher, R., A. Perelygin, J. Wu, J. Chuang, C. D. Manning, A. Y. Ng and C. Potts.
2013. “Recursive deep models for semantic compositionality over a sentiment tree-
bank”. Proceedings of the 2013 Conference on Empirical Methods in Natural Lan-
guage Processing. Stroudsburg, PA: Association for Computational Linguistics.
1631–1642.
Tai, K. S., R. Socher and C. D. Manning. 2015a. “Improved semantic representations
from tree-structured long short-term memory networks”.
<http://arxiv.org/abs/1503.00075>
Wawer, A. 2012. “Mining co-occurrence matrices for SO-PMI paradigm word candi-
dates”. Proceedings of the Student Research Workshop at the 13th Conference of
the European Chapter of the Association for Computational Linguistics (EACL’12
SRW). Avignon: Association for Computational Linguistics. 74–80.
Wawer, A. 2015. “Towards domain-independent opinion target extraction”. IEEE Inter-
national Conference on Data Mining Workshop, ICDMW 2015. Atlantic City, NJ,
November 14–17, 2015. IEEE. 1326–1331. doi:10.1109/ICDMW.2015.255.
Wawer, A. 2016. “OPFI: A tool for opinion finding in Polish”. Proceedings of the Tenth
International Conference on Language Resources and Evaluation (LREC 2016).
Paris,: European Language Resources Association (ELRA).
Wawer, A. and M. Ogrodniczuk. 2017. “Results of the poleval 2017 competition: Sen-
timent analysis shared task”. 8th Language and Technology Conference: Human
Language Technologies as a Challenge for Computer Science and Linguistics.
Wawer, A. and D. Rogozińska. 2012. “How much supervision? Corpus-based lexeme
sentiment estimation”. 2012 IEEE 12th International Conference on Data Mining
Workshops (ICDMW). 724–730. doi:10.1109/ICDMW.2012.119
Wilson, T., P. Hoffmann, S. Somasundaran, J. Kessler, J. Wiebe, Y. Choi, C. Cardie, E.
Riloff and S. Patwardhan. 2005. “OpinionFinder: A system for subjectivity analy-
sis”. Proceedings of HLT/EMNLP on Interactive Demonstrations (HLT-Demo ’05),
Stroudsburg, PA: Association for Computational Linguistics. 34–35.
468 A. Wawer