BiVaSE: A bilingual variational sentence encoder with randomly initialized Transformer layers
BiVaSE: A bilingual variational sentence encoder with randomly initialized Transformer layers
Author(s): Bence NyékiSubject(s): Morphology
Published by: Akadémiai Kiadó
Keywords: sentence embedding; variational autoencoder; Transformer; pre-training; pooling
Summary/Abstract: Transformer-based NLP models have achieved state-of-the-art results in many NLP tasks including text classification and text generation. However, the layers of these models do not output any explicit representations for texts units larger than tokens (e.g. sentences), although such representations are required to perform text classification. Sentence encodings are usually obtained by applying a pooling technique during fine-tuning on a specific task. In this paper, a new sentence encoder is introduced. Relying on an autoencoder architecture, it was trained to learn sentence representations from the very beginning of its training. The model was trained on bilingual data with variational Bayesian inference. Sentence representations were evaluated in downstream and linguistic probing tasks. Although the newly introduced encoder generally performs worse than well-known Transformer-based encoders, the experiments show that it was able to learn to incorporate linguistic information in the sentence representations.
- Issue Year: 69/2022
- Issue No: 4
- Page Range: 450-473
- Page Count: 24
- Language: English