Ajou University repository

Open-Domain News Event Extraction Method Based on BERT
  • HUANG HAIRUI
Citations

SCOPUS

0

Citation Export

Advisor
Tae-Sun Chung
Affiliation
아주대학교 대학원
Department
일반대학원 인공지능학과
Publication Year
2024-02
Publisher
The Graduate School, Ajou University
Keyword
Event extractionNeural topic modelOpen-domain event extraction
Description
학위논문(석사)--인공지능학과,2024. 2
Abstract
News is an important way for people to obtain information, and in the open news environment, the type of news is increasingly diversified and the scale of news is huge, which causes problems such as information overload and redundancy. The open-domain event extraction task aims to identify and extract various types of event information from predefined text, The task is usually based on methods such as pre-training or neural topic modeling. However, there are a number of problems with existing methods. First, Existing pre-trained models suffer from insufficient feature vector extraction and excessively high embedding dimensions. Second, Existing methods are not rich enough in semantics and lack syntactic structural information, resulting in poor readability of results and insufficient extraction accuracy. Therefore, to address these issues, this paper first improves the open-domain event extraction method based on the neural topic model of BERT, and then dynamically in- tegrates semantic and syntactic dependency information to obtain rich semantic and syn- tactic features, in order to further improve the model performance. The main research is as follows: Proposed an improvement method of neural topic modeling based on BERT. First, BERT is used in the coding layer for pre-training to obtain the contextual representation of the feature sequences. Second, the Umap dimensionality reduction method is used to obtain more extensive local and global information, and the joint distribution of variables is combined with the deep hidden variable probabilistic graph model to further optimize the parameter inference learning process. Finally, the self-attention mechanism is introduced to assign weights to different nodes to reduce the influence of noisy data, so that the model can pay attention to the more critical features, and further improve the performance of the open-domain event extraction model. Keywords: Event extraction, Open-domain event extraction, Neural topic model.
Language
eng
URI
https://aurora.ajou.ac.kr/handle/2018.oak/38823
Journal URL
https://dcoll.ajou.ac.kr/dcollection/common/orgView/000000033549
Show full item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Total Views & Downloads

File Download

  • There are no files associated with this item.