TSAR2022 Shared Task on Lexical Simplification for English (en), Spanish (es) and Portuguese (pt) - Datasets and Evaluation scripts
Please look at the website of the Shared Task for more details about the Evaluation Benchmark, Guidelines, Registration Form, etc...
TSAR-2022 Shared-Task website
There is no training dataset for the TSAR-2022 Shared Task.
However, a sample of 10 or 12 instances with gold standard annotations is provided here as a trial/sample dataset.
Format of the files:
- Format of the trial_none and test_none files: Sentence<TAB>ComplexWord
- Format of the trial_gold and test_gold files: Sentence<TAB>ComplexWord<TAB>Annotation1<TAB>Annotation2<TAB>...<TAB>AnnotationN
The trial dataset consists of a set of 10 instances (for English and Portuguese) and 12 instances (for Spanish) of a sentence, a target complex word. The trial_none files contain only the instances and the trial_gold files contain the instances and set of gold annotations.
- /datasets/trial/tsar2022_en_trial_none.tsv
- /datasets/trial/tsar2022_en_trial_gold.tsv
- /datasets/trial/tsar2022_es_trial_none.tsv
- /datasets/trial/tsar2022_es_trial_gold.tsv
- /datasets/trial/tsar2022_pt_trial_none.tsv
- /datasets/trial/tsar2022_pt_trial_gold.tsv
The test_none files (used for the evaluation benchmark) contain the instances with the sentences and target complex words.
-
English test_none dataset (373 instances)
/datasets/test/tsar2022_en_test_none.tsv -
Spanish test_none dataset (368 instances)
/datasets/test/tsar2022_es_test_none.tsv -
Portuguese test_none dataset (374 instances)
/datasets/test/tsar2022_pt_test_none.tsv
The test_gold files contain the sentences, target complex words, and gold annotations
-
English test_gold dataset (373 instances)
/datasets/test/tsar2022_en_test_gold.tsv -
Spanish test_gold dataset (368 instances)
/datasets/test/tsar2022_es_test_gold.tsv -
Portuguese test_gold dataset (374 instances)
/datasets/test/tsar2022_pt_test_gold.tsv
The official results for each language (en, es, and pt) can be found in this directory:
/results/official
The following 10 metrics are reported in the official results:
- MAP@1/Potential@1/Precision@1
- MAP@3
- MAP@5
- MAP@10
- Potential@3
- Potential@5
- Potential@10
- Accuracy@1@top_gold_1
- Accuracy@2@top_gold_1
- Accuracy@3@top_gold_1
The extended results for each language (en, es, and pt) can be found in this directory:
/results/extended
The following metrics are reported in the extended results:
- Potential@K K={1..10}
- MAP@K K={1..10}
- Precision@K K={1..10} (macro-average)
- Recall@K K={1..10} (macro-average)
- Accuracy@K@top_gold_1 K={1..10}
This script evaluates the following metric:
- MAP@1/Potential@1/Precision@1
- MAP@3
- MAP@5
- MAP@10
- Potential@3
- Potential@5
- Potential@10
- Accuracy@1@top_gold_1
- Accuracy@2@top_gold_1
- Accuracy@3@top_gold_1
Script options and help
Evaluation Script for the TSAR-2022 Lexical Simplification Shared Task
Usage: tsar_eval.py <options>
Options:
-h, --help show this help message and exit
--gold_file=<PATH> The path to the file with the gold annotated instances
--predictions_file=<PATH>
The path to file with the predictions
--output_file=<PATH> path to the output file
--verbose Verbose output mode
usage example
python3 ./tsar_eval.py --gold_file ./gold/tsar_es_gold.tsv --predictions_file ./predicted/TEAMNAME_TRACKNAME_RUNNAME.tsv --output_file ./results/TEAMNAME_TRACKNAME_RUNNAME.tsv.eval
A paper describing the compilation of the datasets for English, Portuguese and Spanish that includes several experiments with two state-of-the-art approaches for Lexical Simplification has been published at this link: https://www.frontiersin.org/articles/10.3389/frai.2022.991242
Lexical Simplification Benchmarks for English, Portuguese, and Spanish.
Sanja Štajner, Daniel Ferrés, Matthew Shardlow, Kai North, Marcos Zampieri and Horacio Saggion.
Front. Artif. Intell. Sec. Natural Language Processing.
doi: 10.3389/frai.2022.991242
Preprint available at ArXiV: https://arxiv.org/abs/2209.05301
The python scripts follow AGPL 3.0v license. The datasets (under the /datasets directory) are licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License CC-BY-NC-SA-4.0.