Abstract
We present a neural-encoded mention-hypergraph (named as NEMH in this paper) model for mention-extraction and classification in this paper. Through extraction of textual mention entities, a model is proposed that applies a hypergraph-encoding schema to neural networks. Comparing the results of the proposed model with the previous approaches, the proposed model can thus identify unlimited-length nested mention entities, which is a major milestone in the field. Several experiments are conducted on many datasets used in the baseline approaches, and the obtained results indicated that the designed model has high effectiveness compared to the existing models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Berger, A.L., Pietra, S.A.D., Pietra, V.J.D.: A maximum entropy approach to natural language processing. Comput. Linguist. 22(1), 39–71 (1996)
Baum, L.E., Petrie, T.: Statistical inference for probabilistic functions of finite state Markov chains. Ann. Math. Stat. 37(6), 1554–1563 (1966)
Baum, L.E., Eagon, J.A.: An inequality with applications to statistical estimation for probabilistic functions of Markov processes and to a model for ecology. Bull. Am. Math. Soc. 37(3), 360–363 (1967)
Baum, L.E., Sell, G.R.: Growth transformations for functions on manifolds. Pac. J. Math. 27(2), 211–227 (1968)
Baum, L.E., Petrie, T., Soules, G., Weiss, N.: A maximization technique occurring in the statistical analysis of probabilistic functions of Markov chains. Ann. Math. Stat. 41(1), 164–171 (1970)
Baum, L.E.: An inequality and associated maximization technique in statistical estimation of probabilistic functions of a Markov process. Inequalities 3, 1–8 (1972)
Dyer, C., Ballesteros, M., Ling, W., Matthews, A., Smith, N.A.: Transition based dependency parsing with stack long short term memory. In: Conference on Association for Computational Linguistics, pp. 334–343 (2015)
Florian, R., et al.: A statistical model for multilingual entity detection and tracking. In: Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp. 1–8 (2004)
Finkel, J.R., Manning, C.D.: Nested named entity recognition. In: Conference on Empirical Methods in Natural Language Processing, pp. 141–150 (2009)
Finkel, J.R., Kleeman, A., Manning, C.D.: Efficient, feature-based, conditional random field parsing. In: Conference on Association for Computational Linguistics, pp. 959–967 (2008)
Fine, S., Singer, Y., Tishby, N.: The hierarchical hidden Markov model: analysis and applications. Mach. Learn. 32(1), 41–62 (1998)
Gupta, P., Andrassy, B.: Table filling multi-task recurrent neural network for joint entity and relation extraction. In: International Conference on Computational Linguistics, pp. 2537–2574 (2016)
Guo, S., Chang, M.W., Kiciman, E.: To link or not to link? A study on end-to-end Tweet entity linking. In: Conference on North American Chapter of the Association for Computational Linguistics, pp. 1020–1030 (2013)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Huang, Z., Xu, W., Yu, K.: Bidirectional LSTM-CRF models for sequence tagging. https://arxiv.org/abs/1508.01991 (2015)
Lample, G., Ballesteros, M., Subramanian, S., Kawakami, K., Dyer, C.: Neural architectures for named entity recognition. https://arxiv.org/abs/1603.01360 (2016)
Lu, W., Roth, D.: Joint mention extraction and classification with mention hypergraphs. In: Conference on Empirical Methods in Natural Language Processing, pp. 857–867 (2015)
Lafferty, J.D., Mccallum, A., Pereira, F.C.N.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: International Conference on Machine Learning, pp. 282–289 (2001)
McCallum, A., Freitag, D., Pereira, F.C.N.: Maximum entropy Markov models for information extraction and segmentation. In: International Conference on Machine Learning, pp. 591–598 (1999)
Mintz, M., Bills, S., Snow, R., Jurafsky, D.: Distant supervision for relation extraction without labeled data. In: Conference of ACL-ICJNLP, pp. 1003–1011 (2009)
Muis, A.O., Lu, W.: Labeling gaps between words: recognizing overlapping mentions with mention separators. In: Conference on Empirical Methods in Natural Language Processing, pp. 2598–2608 (2017)
Rosenberg, D.S., Dan, K., Taskar, B.: Mixture-of-parents maximum entropy Markov models. https://arxiv.org/abs/1206.5261 (2012)
Sarawagi, S., Cohen, W.W.: Semi-Markov conditional random fields for information extraction. In: Conference on Neural Information Processing Systems, pp. 1185–1192 (2004)
Zhuo, J., Cao, Y., Zhu, J., Zhang, B., Nie, Z.: Segment-level sequence modeling using gated recursive semi-Markov conditional random fields. In: Conference on Association for Computational Linguistics, pp. 1413–1423 (2016)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Lin, J.CW., Wu, J.MT., Shao, Y., Pirouz, M., Zhang, B. (2019). An Effective BI-encoded Schema for Mention Extraction. In: Lin, JW., Ting, IH., Tang, T., Wang, K. (eds) Multidisciplinary Social Networks Research. MISNC 2019. Communications in Computer and Information Science, vol 1131. Springer, Singapore. https://doi.org/10.1007/978-981-15-1758-7_5
Download citation
DOI: https://doi.org/10.1007/978-981-15-1758-7_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-1757-0
Online ISBN: 978-981-15-1758-7
eBook Packages: Computer ScienceComputer Science (R0)