TY - GEN
T1 - A Comparative Analysis on the Summarization of Legal Texts Using Transformer Models
AU - Núñez-Robinson, Daniel
AU - Talavera-Montalto, Jose
AU - Ugarte, Willy
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - Transformer models have evolved natural language processing tasks in machine learning and set a new standard for the state of the art. Thanks to the self-attention component, these models have achieved significant improvements in text generation tasks (such as extractive and abstractive text summarization). However, research works involving text summarization and the legal domain are still in their infancy, and as such, benchmarks and a comparative analysis of these state of the art models is important for the future of text summarization of this highly specialized task. In order to contribute to these research works, the researchers propose a comparative analysis of different, fine-tuned Transformer models and datasets in order to provide a better understanding of the task at hand and the challenges ahead. The results show that Transformer models have improved upon the text summarization task, however, consistent and generalized learning is a challenge that still exists when training the models with large text dimensions. Finally, after analyzing the correlation between objective results and human opinion, the team concludes that the Recall-Oriented Understudy for Gisting Evaluation (ROUGE) [13] metrics used in the current state of the art are limited and do not reflect the precise quality of a generated summary.
AB - Transformer models have evolved natural language processing tasks in machine learning and set a new standard for the state of the art. Thanks to the self-attention component, these models have achieved significant improvements in text generation tasks (such as extractive and abstractive text summarization). However, research works involving text summarization and the legal domain are still in their infancy, and as such, benchmarks and a comparative analysis of these state of the art models is important for the future of text summarization of this highly specialized task. In order to contribute to these research works, the researchers propose a comparative analysis of different, fine-tuned Transformer models and datasets in order to provide a better understanding of the task at hand and the challenges ahead. The results show that Transformer models have improved upon the text summarization task, however, consistent and generalized learning is a challenge that still exists when training the models with large text dimensions. Finally, after analyzing the correlation between objective results and human opinion, the team concludes that the Recall-Oriented Understudy for Gisting Evaluation (ROUGE) [13] metrics used in the current state of the art are limited and do not reflect the precise quality of a generated summary.
KW - Abstractive text summarization
KW - Benchmark
KW - Deep learning
KW - Natural language processing
KW - Transformers
UR - https://www.scopus.com/pages/publications/85144232675
U2 - 10.1007/978-3-031-20319-0_28
DO - 10.1007/978-3-031-20319-0_28
M3 - Contribución a la conferencia
AN - SCOPUS:85144232675
SN - 9783031203183
T3 - Communications in Computer and Information Science
SP - 372
EP - 386
BT - Advanced Research in Technologies, Information, Innovation and Sustainability - Second International Conference, ARTIIS 2022, Revised Selected Papers
A2 - Guarda, Teresa
A2 - Portela, Filipe
A2 - Augusto, Maria Fernanda
PB - Springer Science and Business Media Deutschland GmbH
T2 - 2nd International Conference on Advanced Research in Technologies, Information, Innovation and Sustainability, ARTIIS 2022
Y2 - 12 September 2022 through 15 September 2022
ER -