Frontiers of Data and Computing ›› 2024, Vol. 6 ›› Issue (6): 62-73.

CSTR: 32002.14.jfdc.CN10-1649/TP.2024.06.007

doi: 10.11871/jfdc.issn.2096-742X.2024.06.007

Previous Articles     Next Articles

Research on the Generation and Evaluation of Judicial Text Summarization Based on Large Language Models

PEI Bingsen(),LI Xin*(),JIANG Zhangtao,LIU Mingshuai   

  1. Information Network Security Academy, People’s Public Security University of China, Beijing 100038, China
  • Received:2024-05-24 Online:2024-12-20 Published:2024-12-20
  • Contact: LI Xin E-mail:597637809@qq.com;lixin@ppsuc.edu.cn

Abstract:

[Objective] With the development of science and technology in modern society, text summarization technology is widely used in various aspects of life and plays a significant role. However, in the judicial field, the use of traditional deep learning models to generate judicial text summaries is facing issues such as redundancy, inconsistent information, and insufficient understanding of text semantics. Moreover, the existing quality evaluation method for generated summaries, ROUGE, is relatively simplistic, focusing only on the overlap between the generated summary and the reference summary without considering deep semantic information. There is a need to construct new multi-dimensional indicators to evaluate the quality of generated summaries. [Methods] This article constructs a vertical large language model in the judicial field with the help of knowledge editing and parameter fine-tuning techniques to generate judicial text summaries, addressing the problem of traditional model’s insufficient understanding of professional judicial texts. It also proposes two evaluation indicators: information missing factor and information density. The large language model is used to extract knowledge from the summaries, measure text knowledge, and calculate the two types of indicators based on the measurement results to further evaluate the quality of the generated summaries at the intrinsic semantic level. [Conclusions] Experiments have proven that the method proposed in this article, which generates judicial text summaries based on a vertical domain large language model, helps to address issues of information redundancy and insufficient understanding of professional texts. Additionally, the two proposed summary evaluation indicators complement the evaluation of summary generation quality from the aspects of information consistency and redundancy, enriching the methods of evaluating summaries.

Key words: large language models, judicial text summary generation, text summary quality evaluation, domain fine-tuning techniques, knowledge extraction