The Dark Side of the Language: Syntax-Based Neural Networks Rivaling Transformers in Unseen Sentences

    October 2023
    Dario Onorati, Leonardo Ranaldi, Aria Nourbakhsh, Arianna Patrizi, Elena Sofia Ruzzetti, Francesca Fallucchi, Francesca Fallucchi, Fabio Massimo Zanzotto
    TLDR Syntax-based neural networks can match Transformers in handling unseen sentences.
    This paper demonstrates that syntax-based neural networks can rival pre-trained Transformers on tasks involving definitely unseen sentences, even after fine-tuning and domain adaptation. Experiments using classification tasks over a DarkNet corpus reveal that syntactic and lexical neural networks perform comparably to pre-trained Transformers. Only with extreme domain adaptation, where BERT is retrained on the test set, do Transformers achieve their usual high performance. Thus, syntax-based models are a viable, more transparent alternative with fewer parameters in scenarios involving truly unseen sentences.
    Discuss this study in the Community →