A Joint Label-Enhanced Representation Based on Pre-trained Model for Charge Prediction

  • Conference paper
  • First Online:
Natural Language Processing and Chinese Computing (NLPCC 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13551))

Abstract

As one of the important subtasks of legal judgment prediction, charge prediction aims to predict the final charge according to the fact description of a legal case. It can help make legal judgments or provide legal professional guidance for non-professionals. Most existing works focus on predicting charges only based on the fact description of a legal case while ignoring the semantic information of charge labels. Moreover, suffering from data imbalance in real applications, they are not applicable to predict few-shot charges by lack of training data. To address these issues, we propose a novel legal text presentation based on pre-trained model for charge prediction, named joint label-enhanced representation (JLER), which provides abundant information of charge labels as additional legal knowledge for pre-trained model to improve the charge prediction performance. JLER can improve predicting accuracy and interpretability by combining the charge label information enhanced by double-layer attention with legal text information, along with relieving the impact of data imbalance by fine-tuning pre-trained model from both text features side and charge label one. Experimental results on two real-world datasets demonstrate that our proposed model achieves significant and consistent improvements compared to the state-of-the-art baselines. Specifically, our model outperforms the baselines by about 13.9% accuracy on few-shot charge prediction. It is indicated that the proposed JLER model has good performance for charge prediction and is prospected to be applied to other subtasks of legal judgement prediction.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
EUR 32.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or Ebook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
EUR 29.95
Price includes VAT (Germany)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
EUR 85.59
Price includes VAT (Germany)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
EUR 106.99
Price includes VAT (Germany)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free ship** worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Nagel, S.S.: Applying correlation analysis to case prediction. Tex. l. Rev. 42(7), 1006–1017 (1964)

    Google Scholar 

  2. Segal, J.A.: Predicting supreme court cases probabilistically: the search and seizure cases, 1962–1981. Am. Political Sci. Rev. 78(4), 891–900 (1984)

    Article  Google Scholar 

  3. Lauderdale, B.E., Clark, T.S.: The supreme court’s many median justices. Am. Political Sci. Rev. 106(4), 847–866 (2012)

    Article  Google Scholar 

  4. Katz, D.M., Bommarito, M.J., Blackman, J.: A general approach for predicting the behavior of the supreme court of the United States. PLoS ONE 12(4), e0174698 (2017)

    Article  Google Scholar 

  5. Hu, Z., Li, X., Tu, C., Liu, Z., Sun., M.: Few-shot charge prediction with discriminative legal attributes. In: Proceedings of the COLING (2018)

    Google Scholar 

  6. Zhong, H., Guo, Z., Tu, C.: Legal judgment prediction via topological learning. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 3540–3549. Association for Computational Linguistics, Brussels, Belgium (2018)

    Google Scholar 

  7. Luo, B., Feng, Y., Xu, J., Zhang, X., Zhao, D.: Learning to predict charges for criminal cases with legal basis, pp. 2727–2736 (2017)

    Google Scholar 

  8. Shaghaghian, S., Feng, L.Y., Jafarpour, B., Pogrebnyakov, N.: Customizing contextualized language models for legal document reviews. In: Proceedings of the IEEE International Conference on Big Data (Big Data), pp. 2139–2148 (2020)

    Google Scholar 

  9. Shao, Y., et al.: BERT-PLI: modeling paragraph-level interactions for legal case retrieval. In: Proceedings of IJCAI, pp. 3501–3507 (2020)

    Google Scholar 

  10. Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. Ar**v abs/1907.11692 (2019)

    Google Scholar 

  11. Zhong, H., Zhang, Z., Liu, Z., Sun, M.: Open Chinese Language Pre-trained Model Zoo. Technical Report (2019)

    Google Scholar 

  12. Chalkidis, I., Fergadiotis, M., Malakasiotis, P., Aletras, N., Androutsopoulos, I.: LEGAL-BERT: “preparing the muppets for court”. In: Proceedings of EMNLP: Findings, pp. 2898–2904 (2020)

    Google Scholar 

  13. Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL HLT, pp. 4171–4186 (2019)

    Google Scholar 

  14. **ao, C., et al.: CAIL2018: a large-scale legal dataset for judgment prediction. Ar**v abs/1807.02478 (2018)

    Google Scholar 

  15. Kim, Y.: Convolutional neural networks for sentence classification. EMNLP (2014)

    Google Scholar 

  16. Salton, G., Buckley, C.: Term-weighting approaches in automatic text retrieval. Inf. Process. Manage. 24(5), 513–523 (1988)

    Article  Google Scholar 

  17. Suykens, J.A., Vandewalle, J.: Least squares support vector machine classifiers. Neural Process. Lett. 9(3), 293–300 (1999)

    Article  Google Scholar 

  18. Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: Proceedings of the International Conference on Machine Learning (ICML), pp. 2048–2057 (2015)

    Google Scholar 

  19. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems, pp. 3111–3119 (2013)

    Google Scholar 

  20. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. ar**v preprint ar**v:1412.6980 (2014)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to **gpei Dan .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Dan, J., Liao, X., Xu, L., Hu, W., Zhang, T. (2022). A Joint Label-Enhanced Representation Based on Pre-trained Model for Charge Prediction. In: Lu, W., Huang, S., Hong, Y., Zhou, X. (eds) Natural Language Processing and Chinese Computing. NLPCC 2022. Lecture Notes in Computer Science(), vol 13551. Springer, Cham. https://doi.org/10.1007/978-3-031-17120-8_54

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-17120-8_54

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-17119-2

  • Online ISBN: 978-3-031-17120-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics

Navigation