With the development of Internet big data, how people obtain a large amount of news text information and automatically obtain the information they want from the text information is an urgent task. In order to structure and analyze a large amount of Chinese text information on the Internet, this paper proposes an entity extraction method based on the BERT pre-training model and BiLSTM with the Attention Mechanism. Aiming at the problem that the BiLSTM model can only obtain feature information at the sentence context level, but cannot obtain local feature information. In this paper, based on the BiLSTM model, a BERT feature extraction model is added to obtain word vectors containing contextual semantic information, thereby capturing global and local information. At the same time, an Attention Mechanism is added to improve the effect of the model. The model was trained on the 2018 Football World Cup dataset corpus, and it was verified that the precision, F1 value and recall rate of the model have significantly improved performance on the dataset.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.