Zhang, G. and Nulty, Paul and Lillis, D. (2022) Enhancing legal argument mining with domain pre-training and neural networks. Journal of Data Mining & Digital Humanities NLP4DH , ISSN 2416-5999.
|
Text
JDMDH_submission.pdf - Published Version of Record Available under License Creative Commons Attribution. Download (311kB) | Preview |
Abstract
The contextual word embedding model, BERT, has proved its ability on downstream tasks with limited quantities of annotated data. BERT and its variants help to reduce the burden of complex annotation work in many interdisciplinary research areas, for example, legal argument mining in digital humanities. Argument mining aims to develop text analysis tools that can automatically retrieve arguments and identify relationships between argumentation clauses. Since argumentation is one of the key aspects of case law, argument mining tools for legal texts are applicable to both academic and non-academic legal research. Domain-specific BERT variants (pre-trained with corpora from a particular background) have also achieved strong performance in many tasks. To our knowledge, previous machine learning studies of argument mining on judicial case law still heavily rely on statistical models. In this paper, we provide a broad study of both classic and contextual embedding models and their performance on practical case law from the European Court of Human Rights (ECHR). During our study, we also explore a number of neural networks when being combined with different embeddings. Our experiments provide a comprehensive overview of a variety of approaches to the legal argument mining task. We conclude that domain pre-trained transformer models have great potential in this area, although traditional embeddings can also achieve strong performance when combined with additional neural network layers.
Metadata
Item Type: | Article |
---|---|
School: | Birkbeck Faculties and Schools > Faculty of Science > School of Computing and Mathematical Sciences |
Depositing User: | Paul Nulty |
Date Deposited: | 23 Nov 2022 13:34 |
Last Modified: | 14 Feb 2024 11:56 |
URI: | https://eprints.bbk.ac.uk/id/eprint/49949 |
Statistics
Additional statistics are available via IRStats2.