MedNER: A Service-Oriented Framework for Chinese Medical Named-Entity Recognition with Real-World Application
Abstract
:1. Introduction
1.1. Background and Motivation
1.2. Research Questions and Contributions
- Introduce MedNER, a novel systematic framework featuring a service-oriented architecture that leverages the concept of “Deep Learning as a Service” (DLaaS). This framework is designed to encapsulate multiple pre-trained NER models, providing flexibility, scalability, and ease of integration with existing healthcare IT systems.
- Employ a systematic method for integrating and evaluating multiple pre-trained NER models. This method enhances the framework’s adaptability to diverse datasets and contexts, ensuring robust performance and facilitating automatic comprehensive data analysis. The paper provides a thorough evaluation of the framework using real-world datasets and involving a healthcare expert testing the system.
- Demonstrate an intuitive, user-friendly interface tailored for non-IT experts such as clinicians and healthcare administrators. This design facilitates the efficient use of advanced NER capabilities for clinical decision support and data analysis without requiring extensive technical knowledge.
2. Related Work
3. Proposed Framework
3.1. MedNER: A Service-Oriented Framework
- Text Import Service: the entry point for the MedNER framework, responsible for ingesting and preprocessing various types of clinical texts. This service supports multiple input formats, including plain text, XML, JSON, and common medical document formats used in electronic health records (EHRs). Key functionalities include data cleaning, normalization, format detection, and language detection. In general, this service encompasses (complete or partial) Extraction, Load, and Transformation (ETL) functions that prepare the data for the following tasks in the workflow.
- Model Selection Service: a service that provides users with the ability to choose from a range of pre-trained NER models optimized for various medical domains and tasks. This service addresses the challenge of selecting the most suitable model for specific datasets and contexts. Key features include a model repository of pre-trained models (including those specifically trained on Chinese medical texts), model comparison (for comparing the performance of different models based on various metrics such as accuracy, precision, recall, and F1-score [29]), and automated recommendation (providing a default model for the best-suited models based on the characteristics of the input text and the specific NER task). The pre-trained model repository is structured to facilitate easy access, management, and integration of various pre-trained NER models. It includes model metadata in a structured format like JSON or YAML, describing model architecture, training dataset, performance metrics, compatibility information, and versions. Models are stored in a cloud-based storage system, such as AWS, ensuring scalability and reliability. Access to models is managed through a secure API gateway, allowing authenticated and authorized users or services to fetch models as needed. Model version control could be in place using Git-like repositories.
- Simple Customization Service: a service designed to enhance user-friendliness by allowing non-IT experts to customize the NER process according to their needs. Features include entity selection, enabling users to specify which types of entities (e.g., medical-related entities, general entities like locations and dates, etc.), and model selection, allowing users to select a model with automated parameter settings without requiring deep technical knowledge.
- Deep Learning as a Service (DLaaS): the core component of the MedNER framework, providing the computational infrastructure necessary to run DL models for NER tasks. Key aspects of DLaaS include scalable infrastructure utilizing cloud-based resources, inference using pre-trained models, and performance optimization through techniques such as parallel processing, and GPU acceleration. We will provide more information about this component in the following sections.
- Result and Visualization Service: a service providing users with a comprehensive view of the NER outcomes, facilitating data analysis and decision-making. Key features include user-friendly dashboards and reports summarizing and highlighting the NER results. Multiple export options can be implemented utilizing processing modules to define an interoperable pipeline offering outputs in various formats, such as TXT, CSV, and JSON, to enable a seamless integration of the results with other downstream healthcare information systems and tools.
3.2. Technologies Involved
3.2.1. Traditional NER Model Elements
3.2.2. Deep Learning as a Service
3.2.3. Service-Oriented Architecture (SOA)
- Scalability: Medical texts, especially in large healthcare institutions, can be vast and continuously growing. A service-oriented architecture allows for scalable infrastructure, which can dynamically allocate resources based on the volume of data and the complexity of tasks.
- Flexibility: Different NER tasks may require different models or configurations. By modularizing the system into distinct services, each component can be updated or replaced independently without disrupting the entire framework. This flexibility is crucial for adapting to new advancements in NER technology and varying user needs.
- Resource Efficiency: Utilizing cloud-based resources and techniques such as parallel processing and GPU acceleration ensures that computational resources are used efficiently. This is particularly important for deep learning tasks, which can be resource-intensive.
- User Accessibility: The architecture is designed to be user-friendly, enabling non-IT experts to leverage advanced NER models without needing deep technical knowledge. This democratization of technology ensures broader accessibility and usability across different user groups within the medical field.
3.2.4. User Interface and Experience (UI/UX)
4. Service Layer: Candidate Model Comparison and DLaaS
4.1. Data Preparation
- Disease: Disease entities refer to specific types of diabetes or related complications mentioned in the text, such as Type 2 diabetes, diabetic nephropathy, and so on. This information is critical to understanding the patient’s health status and medical needs.
- Body Part (Anatomy): Body part entities indicate the specific location where the lesion or symptom occurs, e.g., pancreas, blood vessels, etc. In the context of diabetes, clarifying the body part helps in accurate diagnosis and treatment.
- Drug: Drug entities include all medicines used to treat diabetes and its complications, e.g., insulin, metformin, etc. Drug information is extremely important for disease management and drug regimen development.
- Test: Test entities include a variety of tests and measurements used to diagnose and monitor diabetes, including measurement of blood glucose levels and glycosylated hemoglobin (HbA1c). This information plays a critical role in accurately assessing the severity of a diabetic’s condition and tracking progress in treatment.
4.2. Candidate NER Models in Prototype
4.2.1. BiLSTM-CRF
4.2.2. BERT-BiLSTM-CRF
4.3. Evaluation Metrics
4.4. Experiments Configuration and Parameters Tuning
4.5. Comparison Results
5. User Layer: Simple User Interaction
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Jehangir, B.; Radhakrishnan, S.; Agarwal, R. A survey on Named Entity Recognition—Datasets, tools, and methodologies. Nat. Lang. Process. J. 2023, 3, 100017. [Google Scholar] [CrossRef]
- Zhao, W.; Wu, X. Boosting Entity-Aware Image Captioning With Multi-Modal Knowledge Graph. IEEE Trans. Multimed. 2024, 26, 2659–2670. [Google Scholar] [CrossRef]
- Al-Moslmi, T.; Ocaña, M.G.; Opdahl, A.L.; Veres, C. Named Entity Extraction for Knowledge Graphs: A Literature Overview. IEEE Access 2020, 8, 32862–32881. [Google Scholar] [CrossRef]
- Zhang, Y.; Zhang, H. FinBERT–MRC: Financial Named Entity Recognition Using BERT Under the Machine Reading Comprehension Paradigm. Neural Process. Lett. 2023, 55, 7393–7413. [Google Scholar] [CrossRef]
- Wang, C.; Wang, H.; Zhuang, H.; Li, W.; Han, S.; Zhang, H.; Zhuang, L. Chinese medical named entity recognition based on multi-granularity semantic dictionary and multimodal tree. J. Biomed. Inform. 2020, 111, 103583. [Google Scholar] [CrossRef] [PubMed]
- Peng, H.; Zhang, Z.; Liu, D.; Qin, X. Chinese medical entity recognition based on the dual-branch TENER model. BMC Med. Inform. Decis. Mak. 2023, 23, 136. [Google Scholar] [CrossRef]
- Li, J.; Wei, Q.; Ghiasvand, O.; Chen, M.; Lobanov, V.; Weng, C.; Xu, H. A comparative study of pre-trained language models for named entity recognition in clinical trial eligibility criteria from multiple corpora. BMC Med. Inform. Decis. Mak. 2022, 22, 235. [Google Scholar] [CrossRef] [PubMed]
- Ashrafi, I.; Mohammad, M.; Mauree, A.S.; Nijhum, G.M.A.; Karim, R.; Mohammed, N.; Momen, S. Banner: A Cost-Sensitive Contextualized Model for Bangla Named Entity Recognition. IEEE Access 2020, 8, 58206–58226. [Google Scholar] [CrossRef]
- Yu, Y.; Wang, Y.; Mu, J.; Li, W.; Jiao, S.; Wang, Z.; Lv, P.; Zhu, Y. Chinese mineral named entity recognition based on BERT model. Expert Syst. Appl. 2022, 206, 117727. [Google Scholar] [CrossRef]
- Chen, W.; Rabhi, F.; Liao, W.; Al-Qudah, I. Leveraging State-of-the-Art Topic Modeling for News Impact Analysis on Financial Markets: A Comparative Study. Electronics 2023, 12, 2605. [Google Scholar] [CrossRef]
- Nasar, Z.; Jaffry, S.W.; Malik, M.K. Named Entity Recognition and Relation Extraction: State-of-the-Art. ACM Comput. Surv. 2021, 54, 1–39. [Google Scholar] [CrossRef]
- Yu, J.; Yang, X.; Chen, X. A text analysis model based on Probabilistic-KG. In Proceedings of the 2022 IEEE International Conference on Networking, Sensing and Control (ICNSC), Shanghai, China, 15–18 December 2022; pp. 1–6. [Google Scholar]
- Ju, Z.; Wang, J.; Zhu, F. Named Entity Recognition from Biomedical Text Using SVM. In Proceedings of the 2011 5th International Conference on Bioinformatics and Biomedical Engineering, Chongqing, China, 10–12 May 2011; pp. 1–4. [Google Scholar]
- Ahmed, I.; Sathyaraj, R. Named entity recognition by using maximum entropy. Int. J. Database Theory Appl. 2015, 8, 43–50. [Google Scholar] [CrossRef]
- Morwal, S.; Jahan, N.; Chopra, D. Named entity recognition using hidden Markov model (HMM). Int. J. Nat. Lang. Comput. 2012, 1, 4. [Google Scholar] [CrossRef]
- Liu, M.; Tu, Z.; Zhang, T.; Su, T.; Xu, X.; Wang, Z. LTP: A New Active Learning Strategy for CRF-Based Named Entity Recognition. Neural Process. Lett. 2022, 54, 2433–2454. [Google Scholar] [CrossRef]
- Chowdhury, S.; Dong, X.; Qian, L.; Li, X.; Guan, Y.; Yang, J.; Yu, Q. A multitask bi-directional RNN model for named entity recognition on Chinese electronic medical records. BMC Bioinform. 2018, 19, 499. [Google Scholar] [CrossRef] [PubMed]
- Dong, X.; Chowdhury, S.; Qian, L.; Li, X.; Guan, Y.; Yang, J.; Yu, Q. Deep learning for named entity recognition on Chinese electronic medical records: Combining deep transfer learning with multitask bi-directional LSTM RNN. PLoS ONE 2019, 14, e0216046. [Google Scholar] [CrossRef] [PubMed]
- An, Y.; Xia, X.; Chen, X.; Wu, F.-X.; Wang, J. Chinese clinical named entity recognition via multi-head self-attention based BiLSTM-CRF. Artific. Intell. Med. 2022, 127, 102282. [Google Scholar] [CrossRef] [PubMed]
- Sun, C.; Yang, Z.; Wang, L.; Zhang, Y.; Lin, H.; Wang, J. Biomedical named entity recognition using BERT in the machine reading comprehension framework. J. Biomed. Inform. 2021, 118, 103799. [Google Scholar] [CrossRef] [PubMed]
- Wu, G.; Tang, G.; Wang, Z.; Zhang, Z.; Wang, Z. An attention-based BiLSTM-CRF model for Chinese clinic named entity recognition. IEEE Access 2019, 7, 113942–113949. [Google Scholar] [CrossRef]
- Wu, Y.; Huang, J.; Xu, C.; Zheng, H.; Zhang, L.; Wan, J. Research on Named Entity Recognition of Electronic Medical Records Based on RoBERTa and Radical-Level Feature. Wirel. Commun. Mob. Comput. 2021, 2021, 2489754. [Google Scholar] [CrossRef]
- Zhao, P.; Wang, W.; Liu, H.; Han, M. Recognition of the Agricultural Named Entities With Multifeature Fusion Based on ALBERT. IEEE Access 2022, 10, 98936–98943. [Google Scholar] [CrossRef]
- Chai, Z.; Jin, H.; Shi, S.; Zhan, S.; Zhuo, L.; Yang, Y. Hierarchical shared transfer learning for biomedical named entity recognition. BMC Bioinform. 2022, 23, 1–14. [Google Scholar] [CrossRef] [PubMed]
- Brandsen, A.; Verberne, S.; Lambers, K.; Wansleeben, M. Can BERT Dig It? Named Entity Recognition for Information Retrieval in the Archaeology Domain. J. Comput. Cult. Herit. 2022, 15, 1–18. [Google Scholar] [CrossRef]
- Hou, J.; Saad, S.; Omar, N. Enhancing traditional Chinese medical named entity recognition with Dyn-Att Net: A dynamic attention approach. PeerJ Comput. Sci. 2024, 10, e2022. [Google Scholar] [CrossRef]
- Fu, L.; Weng, Z.; Zhang, J.; Xie, H.; Cao, Y. MMBERT: A unified framework for biomedical named entity recognition. Med. Biol. Eng. Comput. 2024, 62, 327–341. [Google Scholar] [CrossRef] [PubMed]
- Chen, P.; Zhang, M.; Yu, X.; Li, S. Named entity recognition of Chinese electronic medical records based on a hybrid neural network and medical MC-BERT. BMC Med. Inform. Decis. Mak. 2022, 22, 315. [Google Scholar] [CrossRef]
- Jiawei, H.; Micheline, K. Data Mining: Concepts and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2006. [Google Scholar]
- Verlinde, P. Error Detecting and Correcting Codes. In Encyclopedia of Information Systems; Bidgoli, H., Ed.; Elsevier: New York, NY, USA, 2003; pp. 203–228. [Google Scholar]
- Bhattacharjee, B.; Boag, S.; Doshi, C.; Dube, P.; Herta, B.; Ishakian, V.; Jayaram, K.R.; Khalaf, R.; Krishna, A.; Li, Y.B.; et al. IBM Deep Learning Service. IBM J. Res. Dev. 2017, 61, 10:1–10:11. [Google Scholar] [CrossRef]
- Chen, W.; El Majzoub, A.; Al-Qudah, I.; Rabhi, F.A. A CEP-driven framework for real-time news impact prediction on financial markets. Serv. Oriented Comput. Appl. 2023, 17, 129–144. [Google Scholar] [CrossRef]
- Ehsan, A.; Abuhaliqa, M.A.M.E.; Catal, C.; Mishra, D. RESTful API Testing Methodologies: Rationale, Challenges, and Solution Directions. Appl. Sci. 2022, 12, 4369. [Google Scholar] [CrossRef]
- Brat. Brat Rapid Annotation Tool. 2022. Available online: https://brat.nlplab.org/ (accessed on 17 July 2024).
- Fu, Y.; Bu, F. Research on Named Entity Recognition Based on ELECTRA and Intelligent Face Image Processing. In Proceedings of the 2021 IEEE International Conference on Emergency Science and Information Technology (ICESIT), Chongqing, China, 22–24 November 2021; pp. 781–786. [Google Scholar]
- Wang, Y.; Sun, Y.; Ma, Z.; Gao, L.; Xu, Y. An ERNIE-based joint model for Chinese named entity recognition. Appl. Sci. 2020, 10, 5711. [Google Scholar] [CrossRef]
- Tavan, E.; Najafi, M. MarSan at SemEval-2022 task 11: Multilingual complex named entity recognition using T5 and transformer encoder. In Proceedings of the 16th international workshop on semantic evaluation (SemEval-2022), Online, 14–15 July 2022; pp. 1639–1647. [Google Scholar]
- Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
- Agrawal, T. Hyperparameter Optimization in Machine Learning: Make Your Machine Learning and Deep Learning Models More Efficient; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
- Ippolito, P.P. Hyperparameter Tuning. In Applied Data Science in Tourism: Interdisciplinary Approaches, Methodologies, and Applications; Egger, R., Ed.; Springer International Publishing: Cham, Switzerland, 2022; pp. 231–251. [Google Scholar]
- James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning; Springer: Berlin/Heidelberg, Germany, 2013; Volume 112. [Google Scholar]
- Takase, T. Dynamic batch size tuning based on stopping criterion for neural network training. Neurocomputing 2021, 429, 1–11. [Google Scholar] [CrossRef]
Model | Description | Advantages | Disadvantages |
---|---|---|---|
BERT [20] | Pre-trained language model based on Transformers, suitable for various NLP tasks. | Handles long-range dependencies and adapts to multiple tasks. | Training and inference are time-consuming and resource-intensive. |
BiLSTM-CRF [21] | Combination of Bidirectional LSTM and Conditional Random Fields for sequence tagging tasks. | Captures contextual information and optimizes sequence labeling. | May be less efficient for processing long sequences. |
RoBERTa [22] | An improved version of BERT with larger training data and longer training duration. | Enhances BERT’s performance, especially in text classification and QA tasks. | High resource requirements. |
ALBERT [23] | Lightweight BERT variant with parameter sharing and factorized embeddings to reduce model size. | Significantly reduces parameter count while maintaining high performance. | Despite fewer parameters, computational complexity remains high. |
XLNet [24] | Utilizes autoregressive and self-attention mechanisms to capture bidirectional context. | Excels in handling long texts. | High training complexity. |
Literature | Model | Key Findings |
---|---|---|
Dong et al. (2019) [18] | Multitask bi-directional RNN | Deep transfer learning is leveraged for the purpose of knowledge transfer and data augmentation in the context of limited data. |
Wang et al. (2020) [5] | MSD_DT_NER | Improved recognition accuracy for Chinese medical texts using multi-granularity semantic dictionary and multimodal tree, combining vocabulary information and position information. |
Li et al. (2022) [7] | BioBERT, BlueBERT, PubMedBERT, SciBERT | A comparative study has been conducted, and PubMedBERT outperformed other pre-trained models in clinical trial eligibility criteria recognition. |
An et al. (2022) [19] | MUSA-BiLSTM-CRF | Multi-head self-attention-based BiLSTM-CRF model with an improved character-level feature representation method combining character embedding and character-label embedding achieved superior performance in Chinese clinical NER tasks. |
Chen et al. (2022) [28] | MC-BERT | Combining BERT with BiLSTM, a CNN, and a CRF layer showed significant improvements in recognizing medical entities in Chinese electronic medical records. |
Peng et al. (2023) [6] | Dual-branch TENER | TENER divides the NER task into two-branch tasks, focusing on entity boundary and type recognition, integrating medical entity dictionary information and Chinese radicals features for improved performance. TENER achieved the best F1 scores on various datasets. |
Entity Type | Prefixed Label | Non-Prefixed Label | Example |
---|---|---|---|
Disease Name | B-Disease | I-Disease | Diabetes, hyperglycemia |
Body Part Name | B-Anatomy | I-Anatomy | Pancreas, blood vessels |
Drug Name | B-Drug | I-Drug | Insulin, metformin |
Test Name | B-Test | I-Test | Glucose measurement, HbA1c test |
Entity Name (Chinese) | Entity Name (English) | Training Set | Test Set |
---|---|---|---|
测试 | Test | 126,529 | 64,995 |
疾病 | Disease | 107,877 | 46,482 |
身体部位 | Anatomy | 78,664 | 35,544 |
药物 | Drug | 47,313 | 17,800 |
Category | Item | Configuration |
---|---|---|
Hardware | CPU RAM GPU | Apple M1 8 cores 8 GB Apple M1 |
Software | Operating System Python Pytorch | MacOS Sonoma14.2 3.8.18 2.1.0 |
Model Name | Embedding Dimension | Maximum Positional | Embedding Hidden Layer Size |
---|---|---|---|
BERT-base-chinese | 768 | 512 | 768 |
Parameter Name | Parameter Value | Parameter Name |
---|---|---|
Target size | 31 | Target size |
Learning rate | 1 × 10−3 | Learning rate |
Training period | 25 | Training period |
Hidden layer size | 256 | Hidden layer size |
Filled words | <PAD> | Filled words |
Unknown word | <UNK> | Unknown word |
Num | Precision | Recall | F1 |
---|---|---|---|
10 | 69% | 66% | 67% |
20 | 68% | 66% | 67% |
30 | 68% | 65% | 67% |
40 | 68% | 65% | 67% |
50 | 67% | 65% | 66% |
60 | 67% | 65% | 66% |
70 | 68% | 65% | 67% |
80 | 69% | 64% | 66% |
90 | 69% | 65% | 67% |
99 | 68% | 65% | 66% |
Model Name | Entity Name | Precision (%) | Recall Rate (%) | F1 Score (%) |
---|---|---|---|---|
BERT-BiLSTM-CRF | Disease name | 88 | 90 | 89 |
Body part name | 82 | 85 | 84 | |
Drug name | 84 | 85 | 85 | |
Test name | 84 | 88 | 86 | |
Weighted average | 77 | 79 | 78 | |
BiLSTM-CRF | Disease name | 80 | 78 | 79 |
Body part name | 55 | 56 | 56 | |
Drug name | 71 | 70 | 71 | |
Test name | 77 | 74 | 76 | |
Weighted average | 68 | 66 | 67 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chen, W.; Qiu, P.; Cauteruccio, F. MedNER: A Service-Oriented Framework for Chinese Medical Named-Entity Recognition with Real-World Application. Big Data Cogn. Comput. 2024, 8, 86. https://doi.org/10.3390/bdcc8080086
Chen W, Qiu P, Cauteruccio F. MedNER: A Service-Oriented Framework for Chinese Medical Named-Entity Recognition with Real-World Application. Big Data and Cognitive Computing. 2024; 8(8):86. https://doi.org/10.3390/bdcc8080086
Chicago/Turabian StyleChen, Weisi, Pengxiang Qiu, and Francesco Cauteruccio. 2024. "MedNER: A Service-Oriented Framework for Chinese Medical Named-Entity Recognition with Real-World Application" Big Data and Cognitive Computing 8, no. 8: 86. https://doi.org/10.3390/bdcc8080086
APA StyleChen, W., Qiu, P., & Cauteruccio, F. (2024). MedNER: A Service-Oriented Framework for Chinese Medical Named-Entity Recognition with Real-World Application. Big Data and Cognitive Computing, 8(8), 86. https://doi.org/10.3390/bdcc8080086