Professional Documents
Culture Documents
Thirumaligai Eisner 4 Ab 1
Thirumaligai Eisner 4 Ab 1
Annotated Bibliography
Information, vol. 14, no. 5, 2023, p. 271. ProQuest Central Student; Publicly Available
quantifying the dissimilarity between texts. The article primarily compared the
embeddings created using the all-MiniLM-L6-v2 model, finding that the Jensen-
Shannon divergence performed very strongly across all 3 tasks, which are
clustering texts by author, subject, and time period, with the vector embeddings
also performing very well, while the Jaccard Distance wasn't as effective,. The
article discusses language models on a very high level, having a higher barrier to
entry than the other articles I have reviewed. The information contained,
used to fine-tune the models used. However, while the information contained is
very useful, I wouldn't annotate this article, as it's too advanced for the moment.
Son, Jungha, and Boyoung Kim. "Translation Performance from the User's Perspective of Large
Language Models and Neural Machine Translation Systems." Information, vol. 14, no.
10, 2023, p. 574. ProQuest Central Student; Publicly Available Content Database,
https://doi.org/10.3390/info14100574.
2
by Jungha Son and Boyoung Kim compares and contrasts the language translation
annotate the parts on the different metrics used to compare the models, which are
their scores in the BLEU, chrF, and TER metrics, as well as their performance in
translating specific language pairs, to understand how Language Models are used
in language translation, as well as how they are graded and scored in their
capabilities. Skimming through the article, I could note down and utilize the
for my Senior Project, before diving into higher-level knowledge such as specific
Zhu, Wenhao, et al. "Multilingual Machine Translation with Large Language Models: Empirical
www.proquest.com/working-papers/multilingual-machine-translation-with-large/
docview/2799277250/se-2?accountid=41498.
Published on April 10, 2023, with the current version revised as of October 29,
Empirical Results and Analysis" by Wenhao Zhu and others discusses 2 primary
3
questions, which are, "1) How LLMs perform MMT over massive languages?"
and "2) Which factors affect the performance of LLMs?" The article initially
and LLaMA2-7B, for translating from English into other languages, comparing
their performances with those of Google Translate, and concluding that general
LLMs still have a long way to go for translation compared to the most common
LLM for translation, Google Translate. The article continues with the second
areas, while also showing their strengths and weaknesses in translation. I will
annotate this source, as it will prove incredibly useful for the foundations of my