نبذة مختصرة : Using long short-term memory (LSTM) networks to build a named entity recognition model is important for the task of named entity recognition. However, traditional memory networks lack a direct connection between input information and hidden states, leading to key feature information not being fully learned during training and causing information loss. This paper designs a bidirectional variant of the long short-term memory (BiLSTM) network called Mogrifier-BiGRU, which combines the BERT pre-trained model and the conditional random field (CRF) network model. The Mogrifier gating interaction unit is set with more hyperparameters to achieve deep interaction of gating information, changing the relationship between input and hidden states so that they are no longer independent. By introducing more nonlinear transformations, the model can learn more complex input–output mapping relationships. Then, by combining Bayesian optimization with the improved Mogrifier-BiGRU network, the optimal hyperparameters of the model are automatically calculated. Experimental results show that the model method based on the gating interaction mechanism can effectively combine feature information, improving the accuracy of Chinese-named entity recognition. On the dataset, an F1-score of 85.42% was achieved, which is 7% higher than traditional methods and 10% higher for the accuracy of some entity recognition.
No Comments.