ScholarMate
客服热线:400-1616-289

BERTBooster: A knowledge enhancement method jointing incremental training and gradient optimization

Jiang, Wenchao; Lu, Jiarong; Liang, Tiancai; Hong, Xiao*; Lu, Jianfeng; Wu, Tao
Science Citation Index Expanded
广东工业大学; 广州大学

摘要

The knowledge-enhanced BERT model solves the problem of lacking knowledge in downstream tasks by injecting external expertize, and achieves higher accuracy compared with BERT model. However, owning to large-scale external knowledge is utilized into knowledge-enhanced BERT, some shortcomings comes such as information noise, lower accuracy and weak generalization ability, and so on. To solve this problem, a knowledge enhancement method BERTBooster which combines incremental learning and gradient optimization is proposed. BERTBooster disassembles the input text corpus into entity noun sets through entity noun recognition, and uses the incremental learning task denoising entity auto-encoder to create an incremental task set of entity nouns and external knowledge triples. Furthermore, BERTBooster introduces a new gradient optimization algorithm ChildTuningF into BERT model to improve the generalization ability. BERTBooster can effectively improve the factual knowledge cognition ability of CAGBERT model and improve the accuracy of the model in downstream tasks. Experiments are carried out on six public data sets such as Book_Review, LCQMC, XNLI, Law_QA, Insureace_QA, and NLPCC-DBQA. The experimental results show that the accuracy rate in downstream tasks is increased by 0.65% on average after using BERTBooster on CAGBERT.

关键词

gradient optimization incremental training knowledge enhanced natural language processing pretraining models