Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational AutoEncoders

Yingji Zhang, Marco Valentino, Danilo Carvalho, Ian Pratt-Hartmann, Andre Freitas

Research output: Chapter in Book/Conference proceedingConference contributionpeer-review

1 Downloads (Pure)

Abstract

The injection of syntactic information in Variational AutoEncoders (VAEs) can result in an overall improvement of performances and generalisation. An effective strategy to achieve such a goal is to separate the encoding of distributional semantic features and syntactic structures into heterogeneous latent spaces via multi-task learning or dual encoder architectures. However, existing works employing such techniques are limited to LSTM-based VAEs. This work investigates latent space separation methods for structural syntactic injection in Transformer-based VAE architectures (i.e., Optimus) through the integration of graph-based models. Our empirical evaluation reveals that the proposed end-to-end VAE architecture can improve theoverall organisation of the latent space, alleviating the information loss occurring in standard VAE setups, and resulting in enhanced performances on language modelling and downstream generation tasks.
Original languageEnglish
Title of host publicationFindings of the Association for Computational Linguistics: NAACL 2024
PublisherAssociation for Computational Linguistics (ACL)
Pages474-489
DOIs
Publication statusE-pub ahead of print - 16 Jun 2024
EventFindings of the Association for Computational Linguistics: NAACL 2024 - Mexico City, Mexico
Duration: 1 Jun 20241 Jun 2024

Conference

ConferenceFindings of the Association for Computational Linguistics: NAACL 2024
Period1/06/241/06/24

Fingerprint

Dive into the research topics of 'Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational AutoEncoders'. Together they form a unique fingerprint.

Cite this