by Osma Suominen, Ilkka Koskenniemi
Introduction
Libraries and other cultural heritage institutions are increasingly adopting natural language processing (NLP) and machine learning technologies to assist in the production of metadata. A common need in such settings is to automatically or semi-automatically classify text documents, for example performing subject indexing or to classify documents into mutually exclusive classes. One practical tool for these kinds of tasks is the automated subject indexing toolkit Annif (Suominen, Inkinen and Lehtinen 2022; Suominen 2019).
This study concentrates on a small but nevertheless important aspect of automated text classification systems, namely the use of different preprocessing and normalization methods such as stemming and lemmatization. These are particularly important in the traditional NLP setting, where text classification pipelines commonly include text preprocessing (e.g. tokenization and word normalization such as stemming or lemmatization), vectorization (turning words into numbers, often represented as vectors using a tf-idf transformation), model training and inference. The goal of word normalization in such a pipeline is to reduce the amount of unimportant linguistic variation by excluding aspects such as singular vs. plural and inflected word forms. For example, the words eat, ate, eats, eating and eaten all share the same base form eat and refer to the same activity, yet without normalization they would all be considered completely different words and a classification algorithm would need to learn their meanings separately. Normalizing them all into the same form with a stemming or lemmatization algorithm reduces the amount of noise in the data and will in many cases both reduce the required amount of training data and improve overall classification results.
Most of the algorithms implemented in Annif follow this traditional NLP paradigm, and in the modular architecture of Annif, the text preprocessing is performed by an analyzer module. This study explores the effect of different text normalization choices on the quality of the resulting automated subject indexing suggestions. The comparison includes both algorithms currently implemented as analyzer modules in Annif as well as several other lemmatization algorithms. We mainly selected open source lemmatization tools that were easy to use from Python code. As this study was performed in a collaborative project between the National Library of Finland and Lingsoft Inc., we also included the proprietary Lingsoft Language Management Central (LMC) platform for language structure analysis in the comparison.
In order to compare the effect of different analyzer methods, we set up a series of experiments where the setting (corpora, algorithms, parameter settings, evaluation metrics) is static except for the text normalization method. We chose two subject prediction tasks, informed by the requirements of the Finto AI automated subject indexing service. The first task is an example of subject indexing with a thesaurus, where the goal is to assign a small number (typically 5–12) of relevant subjects to each document, which will improve the discoverability of the collection. The second task is an example of library classification, where the goal is to place each document in a single class or category – which, in a traditional library setting, would determine the location of a book on the shelf. We selected appropriate corpora, algorithms and other settings for each of the two tasks:
- Automated subject indexing with the General Finnish Ontology YSO (including YSO Places). By now, this is a well established task which has been supported by the Finto AI service, and its experimental predecessor systems, for several years. This task is further divided by language: Finnish, Swedish and English.
- Automated classification with the Finnish Public Library Classification YKL, which is closely related to the Dewey Decimal Classification used around the world. This is currently a more experimental task which, at the time of writing, has not yet been turned into a production service. However, we have collected suitable training and evaluation data and tested different algorithms on this task. We only perform this task in Finnish due to the lack of enough training data in other languages.
The goal of these experiments is to find out what kind of lemmatization methods (or other kinds of text normalization) give the best results when used in a typical NLP pipeline that performs text classification. The experiments are not intended to benchmark the classification algorithms themselves, as this has been done before (see, e.g., Suominen, Inkinen and Lehtinen 2022; Uhlmann and Grote 2021), but to highlight the effect of the lemmatization method on the downstream evaluation results. We hope that the experiment will help inform choices about which lemmatization method to select for a particular type of task and classification setting and to quantify how much that choice matters for the success of the classification task. Although the experimental setting is focused on the Annif tool, the results should generalize to other text classification settings that follow the traditional NLP paradigm, for example custom text classifiers implemented with the scikit-learn machine learning toolkit.
The experiment compares a wide range of lemmatization methods, ranging from very simplistic rule-based approaches to computationally intensive state-of-the-art neural network pipelines which perform many other functions than lemmatization, including part-of-speech (POS) tagging, dependency resolution, named entity recognition (NER) and named entity linking (NEL). For this experiment, only the result of the lemmatization matters, and it doesn’t even have to be linguistically correct, just consistent enough to adequately support the classification algorithms that make use of the lemmatized output. As an example, simple stemming algorithms convert words into stems (e.g. “qualities” becomes “qualiti” using Snowball stemming) that often aren’t valid word forms but can still be used to group together related words.
Corpora for training and evaluation
We selected preexisting corpora that we had already been using to train and evaluate Annif models, in particular in the Finto AI service for automated subject indexing. The corpora consist of manually indexed raw text in the form of TXT and TSV files. Table 1 gives an overview of the corpora used in the experiments. The total amount of raw text included in all the corpora was approximately 1.1 GB.
Task / language | Data set | Train set size | Used for training | Test set size | Raw text |
---|---|---|---|---|---|
YSO Finnish | finna | 7.0M records | Parabel, Bonsai | – | 360 MB |
jyu-theses | 500 documents | MLLM | 764 documents | 270 MB | |
bookdesc | 500 documents | MLLM | 1311 documents | 1.9 MB | |
YSO Swedish | finna | 920k records | Parabel, Bonsai | – | 50 MB |
jyu-theses | 103 documents | MLLM | 32 documents | 25 MB | |
bookdesc | 744 documents | MLLM | 560 documents | 1.1 MB | |
YSO English | finna | 2.0M records | Parabel, Bonsai | – | 120 MB |
jyu-theses | 498 documents | MLLM | 300 documents | 190 MB | |
bookdesc | 826 documents | MLLM | 447 documents | 1.3 MB | |
YKL Finnish | bookdesc | 51187 documents | Parabel, Bonsai | 4237 documents | 46 MB |
Table 1. Overview of document corpora used in the experiment.
To test subject indexing using concepts from the General Finnish Ontology YSO, we chose a snapshot of YSO and YSO Places dated 2021-09-23, training corpora based on metadata records collected from the Finna.fi discovery system (finna), and two sets of manually indexed full text documents: Master’s and doctoral theses from the University of Jyväskylä (jyu-theses), as well as titles and short descriptions of non-fiction books collected from the database of Kirjavälitys Oy (bookdesc). These corpora include documents in three languages (Finnish, Swedish and English) and have been used to train and evaluate the models in the Annif-based automated subject indexing service Finto AI.
We also included a Finnish language corpus which is based on the same book descriptions as the bookdesc data set, but classified using the Finnish Public Library Classification YKL, in order to test the effect of text preprocessing strategies in an automated classification setting.
Text preprocessing methods
We included nine text preprocessing methods in the comparison: the baseline methods simple and snowball as well as the more advanced lemmatization methods voikko, spacy, simplemma, udpipe, stanza, lmc and tnpp. Not all methods work on all three languages, as detailed below. The first four methods are implemented in Annif (version 0.57.0 was used for the experiments), so we relied on the built-in support. For the other methods, we performed the lemmatization of the data files outside the Annif tool, in most cases using Jupyter notebooks (Google Colaboratory in the case of TNPP and a custom Python script in the case of LMC), and stored the results in corpus files that were subsequently used instead of the original, unlemmatized files. We configured the corresponding Annif projects to use the simple analyzer, which converts the input text into lower case.
Baseline methods
The simple analyzer, implemented in Annif, is a baseline, language-agnostic method that folds the input text into lower case but does not attempt to perform any stemming or lemmatization.
The snowball analyzer in Annif is another baseline method that converts words into lower case and additionally performs stemming using the Snowball algorithm, as implemented in the Python Natural Language Toolkit (NLTK). The stemming rules are specific to each language and the implementation in NLTK supports at least 16 languages, including Finnish, Swedish and English.
Lemmatization methods
The voikko analyzer in Annif is based on the Voikko morphological analysis tool, which was originally developed as a Finnish language spell checker for the OpenOffice and LibreOffice applications. This analyzer also converts individual words into lower case. Voikko only supports the Finnish language.
The spacy analyzer in Annif performs lemmatization using spaCy (version 3.2.4), which is a popular, multilingual NLP toolkit. Lemmatization requires installing a language-specific pretrained model, but among the languages used in this comparison, only English was supported (support for Finnish and Swedish was added in spaCy 3.3, which was released after the experiments had already been completed). We tested two variations of this method: with and without forced lowercasing of individual lemmas. The lowercased variant is called spacy_lower in the remainder of this paper.
The simplemma method was implemented using the simplemma library (version 0.6.0), a simple multilingual lemmatizer for Python which does not need morphosyntactic information and works on the level of individual words, which makes it extremely fast. It has support for 38 languages, including the three languages used in this comparison. The following code demonstrates how Simplemma was initialized and used for lemmatization:
# Load the English model import simplemma langdata = simplemma.load_data('en') # Lemmatize text using Simplemma def lemmatize(text): lemmas = simplemma.text_lemmatizer(text, langdata) return " ".join(lemmas)
The udpipe method was implemented using the spaCy + UDPipe package (version 1.0.0) that wraps the UDPipe NLP pipeline (version 1, originally implemented in C++ and accessible through Python bindings) so that it can be used via the spaCy API. It supports more than 100 languages or variants, including all three languages used in this comparison; the language-specific models need to be downloaded separately. The following code demonstrates how spaCy + UDPipe was initialized and used for lemmatization:
# Download the English model import spacy_udpipe spacy_udpipe.download('en') # Load the English model nlp = spacy_udpipe.load('en') # Lemmatize text using spaCy + UDPipe def lemmatize(text): lemmas = [token.lemma_ for token in nlp(text)] return " ".join(lemmas)
The stanza method was implemented using the multilingual Stanza NLP toolkit (version 1.3.0) to perform lemmatization. Stanza is based on a neural network architecture and like spaCy and UDPipe, it requires downloading language-specific pretrained models. It supports all three languages that were used in this comparison. The following code demonstrates how Stanza was initialized and used for lemmatization:
# Download the English model import stanza stanza.download('en') # Limit the maximum number of threads for best performance import os os.environ["OMP_NUM_THREADS"] = "1" # Initialize the Stanza NLP module for English # Only include the processors necessary for lemmatization nlp = stanza.Pipeline(lang='en', processors='tokenize,mwt,pos,lemma') # Lemmatize text using Stanza def lemmatize(text): doc = nlp(text) lemmas = (word.lemma for sent in doc.sentences for word in sent.words if word.lemma) return " ".join(lemmas)
The lmc method refers to lemmatization using Lingsoft’s proprietary text analysis and proofing platform, the Language Management Central (LMC). It is a language technology platform through which various NLP tools are made available as cloud services, including lemmatization, POS tagging, syntactical disambiguation, NER and NEL, keywording and sentiment analysis. Supported languages include Finnish, Swedish, English, Norwegian, and Danish. Its lemmatization capabilities are based on Lingsoft’s FINTWOL, a rule-based morphological model of the Finnish language and FINCG, a constraint grammar disambiguator for Finnish. The TWOL and CG for the lemmatization are based on the two-level morphology (Koskenniemi, 1983) and constraint grammar (Karlsson, 1990) research by two of Lingsoft’s founders, professors Fred Karlsson and Kimmo Koskenniemi.
In lmc_yso, a semantic layer extension is also used, which marks the token sequences in running text where it finds labels of terms in a given terminology/ontology like YSO. For example:
“Negatiivinen paine ilmenee pimeässä energiassa, joka työntää galakseja kauemmas toisistaan.” (English translation: Negative pressure occurs in dark energy that pushes the galaxies further apart.)
would be tagged as
“negatiivinen paine ilmetä pimeä energia [pimeä_energia] , joka työntää galaksi kauemmas toinen .” (English translation: Negative pressure occurs in dark energy [dark_energy] that pushes the galaxies further apart.)
It attempts to mark only the sequences where the term’s label appears in a syntactically valid manner (e.g. proper congruence between constituents of a noun phrase). The idea of using the YSO tagging in this test was to attempt to make it easier for the text classification algorithm to distinguish between cases where the different parts of a multi-word term label appear independently (e.g. “LED-valaisin valaisee pimeässä energiaa säästäen.”) in the text from cases where the term is actually being used.
The analysis run was done as a batch run on a desktop Linux computer at Lingsoft’s premises. The various source files were processed using a top-level shell script that sent them to a command-line wrapper tool (a Python script). This tool performed the actual calls to the lmc API. This run utilized a feature in lmc that processes the tab-separated values (TSV) format.
The tnpp method was implemented using the Turku Neural Parser Pipeline (TNPP), a NLP pipeline based on neural networks, to perform lemmatization. Pretrained models were available for Finnish and English, but we chose to test this method only for the YKL classification task because of its high resource requirements, including the need for GPU computing. We used Google Colaboratory to perform the lemmatization of train and test documents, adapting an example notebook (tnpp-diaparse.ipynb dated 2022-01-19) provided by the TNPP project. The notebook provides functionality to parse text and print the parsed structure in the CoNLLU output format. For lemmatization, we used the following function whose main purpose is to convert the CoNLLU format into a string of lemmas:
from tnparser.pipeline import read_pipelines, Pipeline # What pipelines do we have for the Finnish model? available_pipelines = read_pipelines( "models_fi_tdt_dia/pipelines.yaml") # Instantiate one of the pipelines p = Pipeline(available_pipelines["parse_plaintext"]) # Lemmatize text using TNPP def lemmatize(text): parsed = p.parse(text) # parsed string in CoNLLU format lemmas = [] for line in parsed.split("\n"): line = line.strip() if not line or line.startswith("#"): continue flds = line.split("\t") lemmas.append(flds[2]) return ' '.join(lemmas)
Example output of each method
In the following tables, we show the output of the different preprocessing methods when applied on example text. The example texts are titles of books selected from the book description data set. We picked examples that illustrate the differences between the methods; usually the differences in the output of the different methods are more subtle. Note that in these examples, we have retained the original case of words as returned by the normalization algorithms, even though text was subsequently folded to lower case in most of the Annif experiments.
original | Hääd pulmapäeva ja yli 1000 muuta viron kielen riskisanaa |
simple | hääd pulmapäeva ja yli 1000 muuta viron kielen riskisanaa |
snowball | hääd pulmapäev ja yli 1000 muuta viro kiele riskisan |
voikko | Hääd pulmapäeva ja yli 1000 muuttaa viro kieli riskisana |
simplemma | Hääd pulmapäeva ja yli 1000 muu viro kieli riskisana |
udpipe | Hääd pulmapäeva ja yli 1000 muu viro kieli riskisana |
stanza | hääd pulma#päeva ja yli 1000 muu viro kieli riski#sana |
lmc | hääd pulmapäeva ja yli 1000 muuttaa viro kieli riskisana |
lmc_yso | hääd pulmapäeva ja yli 1000 muuttaa viro kieli [viro_kieli] riskisana |
tnpp | hääd pulma#päeva ja yli 1000 muu virko kieli riski#sana |
Table 2. Original Finnish input text and the output of each preprocessing method.
The Finnish input text used as an example in Table 2 is a bit tricky to analyze, since the first two words are actually in Estonian language, which is closely related to Finnish; the book is a collection of Estonian words and phrases that resemble Finnish, but whose meaning is different from what a Finnish speaker would expect. The stanza and tnpp analyzers split compound words such as pulmapäev (Estonian for ‘wedding day’) and riskisana (‘risk word’, i.e., a word that is easily misinterpreted) into their constituent parts separated using a hash character. The lmc_yso analyzer has noticed that the phrase viron kielen matches the YSO concept viron kieli (‘Estonian language’). Some words are wrongly lemmatized: muuta (‘other’) is lemmatized into muuttaa (‘to change’) by voikko and lmc, while tnpp lemmatizes viron (‘Estonian’) into virko which is a nonsensical word.
original | Knut Pipping och etableringen av den moderna sociologin vid Åbo Akademi |
simple | knut pipping och etableringen av den moderna sociologin vid åbo akademi |
snowball | knut pipping och etablering av den mod sociologin vid åbo akademi |
simplemma | Knut pipping och etablering av en modern sociologi vid åbo akademi |
udpipe | Knut Pipping och etablering av en modern sociologi vid Åbo Akademi |
stanza | Knut Pipping och etablering av en modern sociologi vid Åbo Akademi |
lmc | knut pipping och etablering av den modern sociologi vid Åbo akademi |
lmc_yso | knut pipping och etablering av den modern [den_modern] sociologi vid Åbo akademi |
Table 3. Original Swedish input text and the output of each preprocessing method.
The Swedish input text in Table 3 is the title of a book about the history of sociology. Again, lmc_yso has noticed that the phrase den moderna matches the YSO concept det moderna (‘modern’, referring to the modern age). There are no obvious mistakes in the lemmatization, although lmc has chosen a different lemma for den moderna than other lemmatizing methods; simplemma, udpipe and stanza selected en modern (‘a modern’) instead of den modern (‘the modern’). There are also differences in the capitalization of the names Knut Pipping (a person) and Åbo Akademi (a university), although these are not relevant for this study due to the case folding performed later in the process.
original | Health-Related Quality of Life and Functioning in Patients with Spinal Fusion |
simple | health-related quality of life and functioning in patients with spinal fusion |
snowball | health-rel qualiti of life and function in patient with spinal fusion |
spacy | Health – relate Quality of Life and Functioning in patient with Spinal Fusion |
simplemma | Health-relate quality of life and function in patient with spinal fusion |
udpipe | health – related quality of Life and functioning in patients with spinal Fusion |
stanza | health -related quality of life and function in patient with spinal Fusion |
lmc | Health-Relate quality of life and function in patient with spinal fusion |
lmc_yso | Health-Relate quality of life [quality_of_life] and function in patient with spinal fusion |
Table 4. Original English input text and the output of each preprocessing method.
The English input text in Table 4 is the title of a doctoral dissertation in the field of medicine. There are differences in how the different methods handle the compound expression Health-Related, as well as in the capitalization; the original text was given in title case and many methods have preserved at least some of the uppercased words. Again, the lmc_yso method has matched the expression quality of life to the YSO concept with the same label.
Model training and evaluation
We used the Data Version Control (DVC) tool to orchestrate the Annif experiments. All the corpus and vocabulary preparation, training and evaluation steps were specified in a single dvc.yaml configuration file that eventually grew to 979 lines. The dvc repro command was used to run the necessary Annif commands. Results of the evaluation runs were written by the Annif tool into JSON files. Finally, a Jupyter notebook was used to collect the results from the JSON files and, with the help of the Pandas data analysis library, format them into tables that were then copied into the Results section of this paper. The DVC automation made it easy to ensure that all experiment runs were performed consistently and correctly. All experiments (both model training and evaluation) were repeated five times and the results of the evaluations averaged in order to minimize the effect of random fluctuations.
Algorithms for text classification
We chose to include two algorithms supported by Annif in the comparison, in a total of three different configurations:
The MLLM algorithm (Maui-like Lexical Matching) originally implemented in Annif is a lexical automated subject indexing algorithm, inspired by the Maui tool (Medelyan 2009). It relies on matching terms (concept labels) in the indexing vocabulary to terms (words and combinations of words) in the document text, so it is important that the normalized terms from the vocabulary match the normalized terms from the text. Without text normalization, the algorithm doesn’t perform well, as it then cannot match e.g. singular terms that appear in document text to plural form terms in the vocabulary or vice versa.
Omikuji is a reimplementation of a family of efficient tree-based machine learning algorithms for multi-label classification, including Parabel (Prabhu et al. 2018) and Bonsai (Khandagale, Xiao, and Babbar 2020). When used with Annif, it can be configured with settings that constrain the size and shape of the decision tree in a way that emulates Parabel or Bonsai, with Parabel being the default configuration. Both the Parabel and Bonsai variations were used in these experiments, as they have slightly different strengths. These algorithms need to be trained on a large number of manually indexed samples and they learn to recognize correlations between words (or word n-grams, i.e. combinations of consecutive words) that appear in training documents and subjects or classes those documents have been manually assigned, without looking at the terms in the vocabulary at all.
MLLM | Parabel | Bonsai |
---|---|---|
backend=mllm | backend=omikuji transform=limit(5000) |
backend=omikuji transform=limit(5000) cluster_balanced=false cluster_k=100 max_depth=3 ngram=2 min_df=2 |
Table 5. Configuration settings for the Parabel and Bonsai backends used for both YSO and YKL tasks. The settings language, vocab and analyzer have been omitted as they vary depending on the task setup.
The settings of the Annif projects used for the three types of algorithm are shown in Table 5. The MLLM configuration relies entirely on default values. The Parabel configuration is also very simple and mostly relies on default values, except for the transform setting which truncates long documents to the given length (5000 characters), which we have previously found to improve results. The Bonsai configuration is more advanced: in addition to the transform setting for truncating documents, the cluster_balanced, cluster_k and max_depth settings constrain the shape of the decision tree to emulate Bonsai; the ngram=2 setting causes bigrams (combinations of two consecutive words/tokens) to be used as features in addition to unigrams (single words); and the min_df=2 setting requires that n-grams (unigrams or bigrams) must appear in at least two training documents in order to be included as features, thus excluding very rare words and expressions that could add noise and make the model unnecessarily large. These settings represent a simplified version of the models that are either currently used in production in the Finto AI service (MLLM, Bonsai) or have been used in the past (Parabel).
YSO subject indexing task
For the YSO subject indexing task in all three languages, we used the MLLM lexical algorithm as well as both the Parabel and Bonsai variations of the Omikuji algorithm, with the settings shown in Table 5. The MLLM models were trained on the training documents combined from both the jyu-theses and bookdesc corpora. The Parabel and Bonsai models were trained on the finna corpus. We did not use the lmc_yso analyzer with MLLM, as the MLLM algorithm itself identifies YSO concepts within the document text and the additional matching performed by lmc_yso most likely would have interfered with that mechanism.
All three types of YSO subject indexing models were evaluated using a combination of test set documents from both the jyu-theses and bookdesc corpora: 764+1311=2075 documents in Finnish, 32+560=592 documents in Swedish, and 300+447=747 documents in English. We chose two metrics for the evaluation: F1@5 (F1 score – harmonic mean of precision and recall – calculated by comparing the top 5 suggestions of the algorithm to the manually assigned subjects) and nDCG (Normalized Discounted Cumulative Gain), a ranking measure where a higher score indicates that the manually assigned, “correct” subjects are closer to the top of the ranked list of suggestions given by the algorithm.
YKL classification task
For the YKL classification task in Finnish, we used the Parabel and Bonsai variations of the Omikuji algorithm, with the settings shown in Table 5; MLLM was not used for this task, as it is not very well suitable for single-class classification. Both models were trained on the train set consisting of 51187 book descriptions manually classified using YKL. The models were evaluated on the test set of 4237 similarly classified book descriptions. We chose two metrics for the evaluation: Accuracy, which is the proportion of documents for which the top suggestion of the algorithm matches the manually assigned class (equivalent to Precision@1 in this setting where each document is assigned exactly one class), and the same nDCG ranking measure that was also used for the YSO subject indexing task.
Results
YSO subject indexing in Finnish
MLLM algorithm | Parabel algorithm | Bonsai algorithm | ||||
---|---|---|---|---|---|---|
F1@5 | nDCG | F1@5 | nDCG | F1@5 | nDCG | |
simple | 0.221 | 0.290 | 0.415 | 0.563 | 0.454 | 0.615 |
snowball | 0.269 | 0.364 | 0.424 | 0.575 | 0.471 | 0.636 |
voikko | 0.305 | 0.413 | 0.423 | 0.577 | 0.472 | 0.639 |
simplemma | 0.302 | 0.411 | 0.418 | 0.567 | 0.465 | 0.630 |
udpipe | 0.287 | 0.389 | 0.407 | 0.555 | 0.450 | 0.609 |
stanza | 0.301 | 0.412 | 0.397 | 0.542 | 0.466 | 0.631 |
lmc | 0.313 | 0.420 | 0.426 | 0.579 | 0.470 | 0.636 |
lmc_yso | – | – | 0.427 | 0.582 | 0.473 | 0.638 |
Table 6. Evaluation results for Finnish YSO subject indexing using different analyzers.
The results of the Finnish language YSO subject indexing task are shown in Table 6. The differences between analyzers are most apparent with the MLLM lexical algorithm, which performs very poorly with the baseline methods (simple and snowball). The lmc method achieved the highest scores with MLLM. For the Parabel and Bonsai algorithms, differences between analyzers were smaller; the lmc_yso, lmc and voikko analyzers as well as the baseline snowball analyzer all achieved high scores. The results for udpipe and stanza were in many cases worse than those of the baseline methods.
YSO subject indexing in Swedish
MLLM algorithm | Parabel algorithm | Bonsai algorithm | ||||
---|---|---|---|---|---|---|
F1@5 | nDCG | F1@5 | nDCG | F1@5 | nDCG | |
simple | 0.171 | 0.242 | 0.409 | 0.564 | 0.454 | 0.623 |
snowball | 0.208 | 0.300 | 0.425 | 0.583 | 0.464 | 0.633 |
simplemma | 0.207 | 0.297 | 0.423 | 0.579 | 0.458 | 0.625 |
udpipe | 0.204 | 0.292 | 0.408 | 0.565 | 0.438 | 0.601 |
stanza | 0.211 | 0.307 | 0.420 | 0.577 | 0.454 | 0.620 |
lmc | 0.217 | 0.310 | 0.425 | 0.585 | 0.464 | 0.629 |
lmc_yso | – | – | 0.428 | 0.591 | 0.462 | 0.631 |
Table 7. Evaluation results for Swedish YSO subject indexing using different analyzers.
The results of the Swedish language YSO subject indexing task are shown in Table 7. As with Finnish, the lexical MLLM method appears to be most sensitive to the choice of analyzer, with the simple baseline method performing poorly. The lmc and lmc_yso analyzers performed best, but also the snowball analyzer gave surprisingly good results.
YSO subject indexing in English
MLLM algorithm | Parabel algorithm | Bonsai algorithm | ||||
---|---|---|---|---|---|---|
F1@5 | nDCG | F1@5 | nDCG | F1@5 | nDCG | |
simple | 0.235 | 0.321 | 0.417 | 0.571 | 0.500 | 0.678 |
snowball | 0.249 | 0.343 | 0.396 | 0.550 | 0.492 | 0.670 |
spacy | 0.241 | 0.334 | 0.400 | 0.553 | 0.495 | 0.675 |
spacy_lower | 0.242 | 0.337 | 0.405 | 0.555 | 0.496 | 0.675 |
simplemma | 0.243 | 0.336 | 0.407 | 0.559 | 0.493 | 0.674 |
udpipe | 0.235 | 0.325 | 0.397 | 0.546 | 0.476 | 0.656 |
stanza | 0.244 | 0.337 | 0.398 | 0.550 | 0.485 | 0.665 |
lmc | 0.247 | 0.340 | 0.401 | 0.556 | 0.486 | 0.667 |
lmc_yso | – | – | 0.415 | 0.574 | 0.488 | 0.671 |
Table 8. Evaluation results for English YSO subject indexing using different analyzers.
The results of the English language YSO subject indexing task are shown in Table 8. Also in this case, the MLLM lexical algorithm doesn’t perform well with the simple analyzer, but works surprisingly well with the snowball analyzer. For the Parabel and Bonsai algorithms, the simple analyzer gives the best results. The lmc and lmc_yso analyzers are not far behind. In general, it appears that for the English language subject indexing task, the more advanced analyzers for the most part do not produce results that would be better than those of one of the baseline methods.
YKL automated classification in Finnish
Parabel algorithm | Bonsai algorithm | |||
---|---|---|---|---|
Accuracy | nDCG | Accuracy | nDCG | |
simple | 0.622 | 0.725 | 0.612 | 0.724 |
snowball | 0.631 | 0.738 | 0.627 | 0.743 |
voikko | 0.637 | 0.747 | 0.632 | 0.750 |
simplemma | 0.638 | 0.748 | 0.632 | 0.750 |
udpipe | 0.631 | 0.742 | 0.629 | 0.746 |
stanza | 0.634 | 0.752 | 0.639 | 0.763 |
lmc | 0.637 | 0.747 | 0.636 | 0.753 |
lmc_yso | 0.639 | 0.748 | 0.636 | 0.753 |
tnpp | 0.629 | 0.750 | 0.639 | 0.763 |
Table 9. Evaluation results for Finnish YKL classification using different analyzers.
The results of the Finnish language YKL classification task are shown in Table 9. The best results were obtained with the lmc_yso, tnpp and stanza analyzers. As with the Finnish language subject indexing task, the baseline analyzers (simple and snowball) performed worst. The differences between different lemmatizer methods were quite small, with the possible exception of udpipe, which was not much better than the snowball analyzer.
Computational cost
We did not specifically measure the performance of each method in terms of computing resources because the methods were run in different computing environments and we didn’t put much effort into optimizing them. However, it was apparent that some methods were much more resource-intensive than others. We can roughly divide the methods into three categories based on resource usage:
- The lightweight methods include the baseline methods simple and snowball as well as the lemmatizers voikko and simplemma. The lmc analyzer (and its lmc_yso variant) is also quite fast, with the full set of corpora lemmatized in approximately two hours on a desktop PC. With the lightweight preprocessing methods, we can expect the execution time of typical Annif training and evaluation runs to be dominated by other processing steps including text vectorization, model training and prediction.
- The methods with moderate resource requirements include spacy (along with the spacy_lower variant) and udpipe. The spacy analyzer is an order of magnitude more resource-intensive than the lightweight methods, adding several hours to the runtime of the English language training and evaluation runs where it was used. Likewise, udpipe was much slower than the lightweight methods, with a full lemmatization of the corpora taking more than an hour on a high-end server with 48 CPU cores used in parallel.
- The most resource-intensive methods were stanza and tnpp. The Stanza lemmatization of the full set of corpora took several hours with 48 CPU cores used in parallel, while the TNPP analysis of just the small YKL corpus (46 MB of raw text) took six hours on the Google Colab platform using GPU computing. In practice, these two methods are unlikely to be suitable for real time usage due to their high computational cost.
Conclusion
In this comparison of text normalization methods, we expected that more advanced lemmatization methods would result in better downstream quality for text classification tasks compared to simple baseline methods (only case folding and simple rule-based stemming). By and large, this was the result we got, with some notable exceptions: in particular, the baseline methods simple and snowball performed well for the English language subject indexing task whereas even the more advanced lemmatizers performed relatively poorly, rarely scoring above the baseline methods. This may be due to the more analytic nature of the English language, which implies that there are fewer inflected word forms than in a synthetic, heavily inflected language such as Finnish. The clearest benefit of lemmatization over baseline methods was seen for the Finnish language, with Swedish falling somewhere in between Finnish and English.
The lexical algorithm MLLM was clearly more sensitive to the choice of text normalization than the Omikuji variants, which was expected, because MLLM needs to be able to accurately match terms in the vocabulary to terms that appear in the text. The Omikuji algorithms, in contrast, are associative and only concerned with words in the text, so the particular text normalization method makes less of a difference.
The differences between lemmatization methods were relatively small, although lmc and lmc_yso usually ranked at or near the top and udpipe did not perform as well as the other methods. Since the relative performance of different analyzers appears to be different for each task, algorithm and language, it is difficult to predict in advance which method will work best in a particular situation. The best method has to be chosen experimentally, by training and evaluating models using both baseline methods and more advanced lemmatization methods. When using the Annif toolkit, this can be performed by adjusting the analyzer setting of a project; in custom NLP pipelines, the normalization method can be changed by changing the code for text preprocessing.
The study was limited to two types of tasks, corpora from three source systems, nine text normalization methods, three text classification algorithms and three languages. It is by no means a comprehensive comparison of text normalization methods, although even with these options the number of individual evaluation runs was 87 due to combinatorial explosion. This figure was further multiplied by the five repetitions we performed for the purpose of smoothing out random variation, so the full runs took several days on a server with 48 CPU cores. A larger comparison would have required access to training and evaluation corpora in other languages, more normalization methods to test or other possible algorithmic variations. Among lemmatizers, we excluded the UralicNLP toolkit, which we didn’t have time to integrate and whose CC BY-NC-ND license could be problematic; Lemmy, which could have been used for Swedish but usage was not straightforward as all the working examples were for the Danish language; as well as some advanced NLP toolkits that would probably have been extremely resource-heavy, e.g. UDPipe version 2 and SparkNLP.
A core reason for setting up this study was to determine the feasibility of adding additional analyzer modules to the Annif toolkit. Of the lemmatizer methods that were not yet implemented in Annif, we determined the simplemma method to be the most promising, as it is very simple and fast, it supports a large number of languages (currently 38), and the empirical results in this study were rather good even though it was never the top performer for any particular task. We have already implemented an analyzer module based on the Simplemma library and it was included in the Annif 0.58 release. Other analyzer integrations may follow later.
References
Khandagale S, Xiao H, and Babbar R. 2020. Bonsai: Diverse and Shallow Trees for Extreme Multi-Label Classification. Machine Learning [Internet]. [cited 2022 June 22]. 109 (11): 2099–2119. Available from: https://doi.org/10.1007/s10994-020-05888-2.
Medelyan O. 2009. Human-Competitive Automatic Topic Indexing [dissertation], The University of Waikato. Available from: https://researchcommons.waikato.ac.nz/handle/10289/3513.
Prabhu, Y, Kag A, Harsola S, Agrawal R, and Varma M. 2018. Parabel: Partitioned Label Trees for Extreme Classification with Application to Dynamic Search Advertising. In Proceedings of the 2018 World Wide Web Conference, 993–1002. WWW ’18. Lyon, France. Available from: https://doi.org/10.1145/3178876.3185998.
Suominen, O. 2019. Annif: DIY automated subject indexing using multiple algorithms. LIBER Quarterly: The Journal of the Association of European Research Libraries [Internet]. [cited 2022 June 22]. 29 (1), 1–25. Available from: https://doi.org/10.18352/lq.10285
Suominen O, Inkinen J, Lehtinen M. 2022. Annif and Finto AI: Developing and Implementing Automated Subject Indexing. JLIS.it [Internet]. [cited 2022 June 22]. 13 (1), 265–282. Available from: https://doi.org/10.4403/jlis.it-12740
Uhlmann S, Grote C. 2021. Automatic subject indexing with Annif at the German National Library (DNB) [presentation slides]. [cited 2022 June 22]. SWIB21 online conference, December 2021. Available from: https://swib.org/swib21/slides/03-02-uhlmann.pdf
Koskenniemi K. 1983. Two-Level Morphology: A General Computational Model for Word-Form Recognition and Production [dissertation]. (Publications; No. 11). University of Helsinki. Department of General Linguistics. Available from: http://hdl.handle.net/10138/305218
Karlsson F. 1990. Constraint grammar as a framework for parsing running text. In: Karlgren H, editor, Proceedings of the 13th International Conference on Computational Linguistics, volume 3, pp. 168-173, Helsinki, Finland. Available from: https://doi.org/10.3115/991146.991176
About the Authors
Osma Suominen is working as an information systems specialist at the National Library of Finland. He is currently working on automated subject indexing, in particular the Annif tool and the Finto AI service, as well as the publishing of bibliographic data as Linked Data. He is also one of the creators of the Finto.fi thesaurus and ontology service and is leading development of the Skosmos vocabulary browser used in Finto. Osma Suominen earned his doctoral degree at Aalto University while doing research on semantic portals and quality of controlled vocabularies within the FinnONTO series of projects.
Ilkka Koskenniemi is working at Lingsoft as a technical architect and currently in charge of developing Lingsoft’s text analysis and proofing tools and their cloud-based interface LMC. He has a background in computer science at the University of Helsinki where he also worked briefly as a research assistant before starting full-time work at Lingsoft.
Acknowledgements
We thank Filip Ginter and Jenna Kanerva for assisting with TNPP technical issues, Michael Stormbom and Sebastian Andersson for providing background information about LMC, and Juho Inkinen for constructive comments.
Subscribe to comments: For this article | For all articles
Leave a Reply