Word-order Typology in Multilingual BERT: A Case Study in Subordinate-Clause Detection

Research output: Chapter in Book/Conference proceedingConference contributionpeer-review

Abstract

The capabilities and limitations of BERT and similar models are still unclear when it comes to learning syntactic abstractions, in particular across languages. In this paper, we use the task of subordinate-clause detection within and across languages to probe these properties. We show that this task is deceptively simple, with easy gains offset by a long tail of harder cases, and that BERT’s zero-shot performance is dominated by word-order effects, mirroring the SVO/VSO/SOV typology.
Original languageEnglish
Title of host publicationProceedings of the 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
EditorsEkaterina Vylomova, Edoardo Ponti, Ryan Cotterell
Place of PublicationSeattle, Washington
PublisherAssociation for Computational Linguistics
Pages11–21
DOIs
Publication statusPublished - Jul 2022
Externally publishedYes
Event4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP - Seattle, United States
Duration: 14 Jul 2022 → …

Workshop

Workshop4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
Country/TerritoryUnited States
CitySeattle
Period14/07/22 → …

Fingerprint

Dive into the research topics of 'Word-order Typology in Multilingual BERT: A Case Study in Subordinate-Clause Detection'. Together they form a unique fingerprint.

Cite this