@Article{iasc.2021.017021, AUTHOR = {Pu Han, Mingtao Zhang, Jin Shi, Jinming Yang, Xiaoyan Li}, TITLE = {Chinese Q\&A Community Medical Entity Recognition with Character-Level Features and Self-Attention Mechanism}, JOURNAL = {Intelligent Automation \& Soft Computing}, VOLUME = {29}, YEAR = {2021}, NUMBER = {1}, PAGES = {55--72}, URL = {http://www.techscience.com/iasc/v29n1/42542}, ISSN = {2326-005X}, ABSTRACT = {With the rapid development of Internet, the medical Q&A community has become an important channel for people to obtain and share medical and health knowledge. Online medical entity recognition (OMER), as the foundation of medical and health information extraction, has attracted extensive attention of researchers in recent years. In order to further improve the research progress of Chinese OMER, LSTM-Att-Med model is proposed in this paper to capture more external semantic features and important information. First, Word2vec is used to generate the character-level vectors with semantic features on the basis of the unlabeled corpus in the medical domain and open domain respectively. Then, the two character-level vectors are embedded into BiLSTM-CRF as features to construct LSTM-Wiki and LSTM-Med models. Finally, Self-Attention mechanism is introduced into LSTM-Med model, and the performance of the model is validated by using the self-labeled data. The 10-fold cross-validation experiment shows that LSTM-Att-Med with Self-Attention mechanism introduced achieves the best performance and the F-value can be up to 91.66%, which is 0.72% higher than that of BiLSTM-CRF. In addition, the experiment result demonstrates that the improvements of F-value are inconsistent for different corpora based on LSTM-Att-Med. The paper also analyzes the recognition performance and error results of different medical entities.}, DOI = {10.32604/iasc.2021.017021} }