Resolving Legalese: A Multilingual Exploration of Negation Scope Resolution in Legal Documents
Version
Published
Date Issued
2024-05-25
Author(s)
Type
Conference Paper
Language
German
Abstract
Resolving the scope of a negation within a sentence is a challenging NLP task. The complexity of legal texts and the lack of annotated in-domain negation corpora pose challenges for state-of-the-art (SotA) models when performing negation scope resolution on multilingual legal data. Our experiments demonstrate that models pre-trained without legal data underperform in the task of negation scope resolution. Our experiments, using language models exclusively fine-tuned on domains like literary texts and medical data, yield inferior results compared to the outcomes documented in prior cross-domain experiments. We release a new set of annotated court decisions in German, French, and Italian and use it to improve negation scope resolution in both zero-shot and multilingual settings. We achieve token-level F1-scores of up to 86.7% in our zero-shot cross-lingual experiments, where the models are trained on two languages of our legal datasets and evaluated on the third. Our multilingual experiments, where the models were trained on all available negation data and evaluated on our legal datasets, resulted in F1-scores of up to 91.1%.
Subjects
K Law (General)
QA75 Electronic computers. Computer science
Conference
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Submitter
Stürmer, Matthias
Citation apa
Christen, R., Shaitarova, A., Stürmer, M., & Niklaus, J. (2024). Resolving Legalese: A Multilingual Exploration of Negation Scope Resolution in Legal Documents. Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024). https://doi.org/10.48550/arxiv.2309.08695
File(s)![Thumbnail Image]()
Loading...
open access
Name
2024.lrec-main.1220.pdf
License
Attribution-NonCommercial 4.0 International
Version
published
Size
402.39 KB
Format
Adobe PDF
Checksum (MD5)
fcd17a521c85549966fbfaacc1fced49
