Abstract
Recent neural Open Information Extraction (OpenIE) models have improved traditional rule-based systems significantly for Chinese OpenIE tasks. However, these neural models are mainly word-based, suffering from word segmentation errors in Chinese. They utilize dependency information in a shallow way, making multi-hop dependencies hard to capture. This paper proposes a Multi-Grained Dependency Graph Neural Network (MGD-GNN) model to address these problems. MGD-GNN constructs a multi-grained dependency (MGD) graph with dependency edges between words and soft-segment edges between words and characters. Our model makes predictions based on character features while still has word boundary knowledge through word-character soft-segment edges. MGD-GNN updates node representations using a deep graph neural network to fully exploit the topology structure of the MGD graph and capture multi-hop dependencies. Experiments on a large-scale Chinese OpenIE dataset SpanSAOKE shows that our model could alleviate the propagation of word segmentation errors and use dependency information more effectively, giving significant improvements over previous neural OpenIE models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
We express n-ray extraction as tuple and binary extraction as triple in this paper.
References
Angeli, G., Premkumar, M.J.J., Manning, C.D.: Leveraging linguistic structure for open domain information extraction. In: Proceedings of ACL-IJCNLP, pp. 344–354 (2015)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: Proceedings of ICLR (2015)
Banko, M., Cafarella, M.J., Soderland, S., Broadhead, M., Etzioni, O.: Open information extraction from the web. Commun. ACM 51, 68–74 (2008)
Che, W., Feng, Y., Qin, L., Liu, T.: N-ltp: a open-source neural chinese language technology platform with pretrained models. ar**v preprint ar**v:2009.11616 (2020)
Cui, L., Wei, F., Zhou, M.: Neural open information extraction. In: Proceedings of ACL, pp. 407–413 (2018)
Del Corro, L., Gemulla, R.: Clausie: clause-based open information extraction. In: Proceedings of WWW, pp. 355–366 (2013)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, pp. 4171–4186 (2019)
Fader, A., Soderland, S., Etzioni, O.: Identifying relations for open information extraction. In: Proceedings of EMNLP, pp. 1535–1545 (2011)
Fan, A., Gardent, C., Braud, C., Bordes, A.: Using local knowledge graph construction to scale seq2seq models to multi-document inputs. In: Proceedings of EMNLP-IJCNLP, pp. 4186–4196 (2019)
Fang, Y., Sun, S., Gan, Z., Pillai, R., Wang, S., Liu, J.: Hierarchical graph network for multi-hop question answering. In: Proceedings of EMNLP, pp. 8823–8838 (2020)
Gu, J., Lu, Z., Li, H., Li, V.O.: Incorporating copying mechanism in sequence-to-sequence learning. In: Proceedings of ACL, pp. 1631–1640 (2016)
Hamilton, W., Ying, Z., Leskovec, J.: Inductive representation learning on large graphs. In: Advances in neural information processing systems (NIPS), pp. 1024–1034 (2017)
Khot, T., Sabharwal, A., Clark, P.: Answering complex questions using open information extraction. In: Proceedings of ACL, pp. 311–316 (2017)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of ICLR (2015)
Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. In: Proceedings of ICLR (2017)
Kolluru, K., et al.: Imojie: iterative memory-based joint open information extraction. In: Proceedings of ACL, pp. 5871–5886 (2020)
Li, S., Zhao, Z., Hu, R., Li, W., Liu, T., Du, X.: Analogical reasoning on Chinese morphological and semantic relations. In: Proceedings of ACL, pp. 138–143 (2018)
Marcheggiani, D., Titov, I.: Encoding sentences with graph convolutional networks for semantic role labeling. In: Proceedings of EMNLP, pp. 1506–1515 (2017)
Mausam, M.: Open information extraction systems and downstream applications. In: Proceedings of IJCAI, pp. 4074–4077 (2016)
Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems (NIPS), pp. 3111–3119 (2013)
Qiu, L., Zhang, Y.: Zore: a syntax-based system for chinese open relation extraction. In: Proceedings of EMNLP, pp. 1870–1880 (2014)
Ratcliff, J.W., Metzener, D.E.: Pattern-matching-the gestalt approach. Dr Dobbs J. 13, 46 (1988)
Schmitz, M., et al.: Open language learning for information extraction. In: Proceedings of EMNLP-CoNLL, pp. 523–534 (2012)
Schuster, M., Paliwal, K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45, 2673–2681 (1997)
Stanovsky, G., Michael, J., Zettlemoyer, L., Dagan, I.: Supervised open information extraction. In: Proceedings of NAACL-HLT, pp. 885–895 (2018)
Sui, D., Chen, Y., Liu, K., Zhao, J., Liu, S.: Leverage lexical knowledge for chinese named entity recognition via collaborative graph network. In: Proceedings of EMNLP-IJCNLP, pp. 3821–3831 (2019)
Sun, M., Li, X., Wang, X., Fan, M., Feng, Y., Li, P.: Logician: a unified end-to-end neural approach for open-domain information extraction. In: Proceedings of WSDM, pp. 556–564 (2018)
Tseng, Y.H., et al.: Chinese open relation extraction for knowledge acquisition. In: Proceedings of EACL, pp. 12–16 (2014)
Tu, Z., Lu, Z., Liu, Y., Liu, X., Li, H.: Modeling coverage for neural machine translation. In: Proceedings of ACL, pp. 76–85 (2016)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (NIPS), pp. 5998–6008 (2017)
Veličković, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. In: Proceedings of ICLR (2018)
Zeng, D., Liu, K., Lai, S., Zhou, G., Zhao, J.: Relation classification via convolutional deep neural network. In: Proceedings of COLING, pp. 2335–2344 (2014)
Zhan, J., Zhao, H.: Span model for open information extraction on accurate corpus. In: Proceedings of AAAI, pp. 9523–9530 (2020)
Zhang, Y., Yang, J.: Chinese NER using lattice LSTM. In: Proceedings of ACL, pp. 1554–1564 (2018)
Zhang, Y., Qi, P., Manning, C.D.: Graph convolution over pruned dependency trees improves relation extraction. In: Proceedings of EMNLP, pp. 2205–2215 (2018)
Acknowledgements
This work is supported by the NSFC Projects (U1736204, 62006136), grants from Bei**g Academy of Artificial Intelligence (BAAI2019ZD0502) and the Institute for Guo Qiang, Tsinghua University (2019GQB0003).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Lyu, Z., Shi, K., Li, X., Hou, L., Li, J., Song, B. (2021). Multi-Grained Dependency Graph Neural Network for Chinese Open Information Extraction. In: Karlapalem, K., et al. Advances in Knowledge Discovery and Data Mining. PAKDD 2021. Lecture Notes in Computer Science(), vol 12714. Springer, Cham. https://doi.org/10.1007/978-3-030-75768-7_13
Download citation
DOI: https://doi.org/10.1007/978-3-030-75768-7_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-75767-0
Online ISBN: 978-3-030-75768-7
eBook Packages: Computer ScienceComputer Science (R0)