Projects per year
Abstract
The injection of syntactic information in Variational AutoEncoders (VAEs) can result in an overall improvement of performances and generalisation. An effective strategy to achieve such a goal is to separate the encoding of distributional semantic features and syntactic structures into heterogeneous latent spaces via multi-task learning or dual encoder architectures. However, existing works employing such techniques are limited to LSTM-based VAEs. This work investigates latent space separation methods for structural syntactic injection in Transformer-based VAE architectures (i.e., Optimus) through the integration of graph-based models. Our empirical evaluation reveals that the proposed end-to-end VAE architecture can improve theoverall organisation of the latent space, alleviating the information loss occurring in standard VAE setups, and resulting in enhanced performances on language modelling and downstream generation tasks.
Original language | English |
---|---|
Title of host publication | Findings of the Association for Computational Linguistics: NAACL 2024 |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 474-489 |
DOIs | |
Publication status | E-pub ahead of print - 16 Jun 2024 |
Event | Findings of the Association for Computational Linguistics: NAACL 2024 - Mexico City, Mexico Duration: 1 Jun 2024 → 1 Jun 2024 |
Conference
Conference | Findings of the Association for Computational Linguistics: NAACL 2024 |
---|---|
Period | 1/06/24 → 1/06/24 |
Fingerprint
Dive into the research topics of 'Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational AutoEncoders'. Together they form a unique fingerprint.Projects
- 1 Active
-
EnnCore: End-to-End Conceptual Guarding of Neural Architectures
Cordeiro, L. (PI), Brown, G. (CoI), Freitas, A. (CoI), Luján, M. (CoI) & Mustafa, M. (CoI)
1/02/21 → 31/12/25
Project: Research