Back to Search Start Over

A semi-supervised learning approach for semantic parsing boosted by BERT word embedding.

Authors :
Bu, Yanbin
Chen, Ting
Duan, Hongxiu
Liu, Mei
Xue, Yandan
Source :
Journal of Intelligent & Fuzzy Systems. 2024, Vol. 46 Issue 3, p6577-6588. 12p.
Publication Year :
2024

Abstract

In the modern world, structured and semi-structured knowledge bases hold a considerable amount of data. There-fore, people who are familiar with formal query languages should not be the only ones who can efficiently and clearly query them. Semantic Parsing (SP) is converting natural language utterances into formal meaning representations. The paper suggests a model for SP that uses a novel method of utilizing the Semi-Supervised Generative Adversarial Network (SS-GAN) to enhance the classifier performance. The proposed SS-GAN extends the fine-tuning of word embedding architectures using unlabeled examples in a generative adversarial environment. We provide a regularization strategy for addressing the mode missing problem and unstable training in SS-GAN. The main viewpoint is to use the extracted feature vectors from the discriminator. Hence, the generator produces outputs by aiding the discriminator's learned features. A reconstruction loss is added to the loss function of the SS-GAN to drive the genera-tor to reconstruct outputs from the discriminator's features, hence steering the generator toward actual data configurations. The proposed reconstruction loss improves the performance of SS-GAN, produces high-quality outputs, and may be combined with other regularization loss functions to improve the performance of diverse GANs. We employ BERT word embedding for our model, which can be included in a downstream task and fine-tuned as a model, while the pre-trained BERT model can capture various linguistic properties. We examine the suggested model using the WikiSQL and SparC datasets, and the analysis findings reveal our model outperforms its rivals. The findings from our experiments indicate that the need for labeled samples can be minimized, down to as few as 100 instances, while still achieving commendable classification outcomes. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10641246
Volume :
46
Issue :
3
Database :
Academic Search Index
Journal :
Journal of Intelligent & Fuzzy Systems
Publication Type :
Academic Journal
Accession number :
176366323
Full Text :
https://doi.org/10.3233/JIFS-233212