Abstract
The capabilities and limitations of BERT and similar models are still unclear when it comes to learning syntactic abstractions, in particular across languages. In this paper, we use the task of subordinate-clause detection within and across languages to probe these properties. We show that this task is deceptively simple, with easy gains offset by a long tail of harder cases, and that BERT’s zero-shot performance is dominated by word-order effects, mirroring the SVO/VSO/SOV typology.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP |
| Editors | Ekaterina Vylomova, Edoardo Ponti, Ryan Cotterell |
| Place of Publication | Seattle, Washington |
| Publisher | Association for Computational Linguistics |
| Pages | 11–21 |
| DOIs | |
| Publication status | Published - Jul 2022 |
| Externally published | Yes |
| Event | 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP - Seattle, United States Duration: 14 Jul 2022 → … |
Workshop
| Workshop | 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP |
|---|---|
| Country/Territory | United States |
| City | Seattle |
| Period | 14/07/22 → … |