Abstract
The knowledge encapsulated in a model is the core factor determining its final performance on downstream tasks. Much research in NLP has focused on efficient methods for storing and adapting different types of knowledge, e.g., in dedicated modularized structures, and on how to effectively combine these, e.g., by learning additional parameters. However, given the many possible options, a thorough understanding of the mechanisms involved in these compositions is missing, and hence it remains unclear which strategies to utilize. To address this research gap, we propose a novel framework for zero-shot module composition, which encompasses existing and some novel variations for selecting, weighting, and combining parameter modules under a single unified notion. Focusing on the scenario of domain knowledge and adapter layers, our framework provides a systematic unification of concepts, allowing us to conduct the first comprehensive benchmarking study of various zero-shot knowledge composition strategies. In particular, we test two module combination methods and five selection and weighting strategies for their effectiveness and efficiency in an extensive experimental setup. Our results highlight the efficacy of ensembling but also hint at the power of simple though often-ignored weighting methods. Further in-depth analyses allow us to understand the role of weighting vs. top-k selection, and show that, to a certain extent, the performance of adapter composition can even be predicted.
Citation
Holtermann Carolin,
Markus
Frohmann,
Navid
Rekab-saz,
Anne Lauscher
What the Weight?! A Unified Framework for Zero-Shot Knowledge Composition
Findings of the Association for Computational Linguistics: EACL 2024,
18th:
1138–1157, 2024.
BibTeX
@inproceedings{Carolin2024WhatTheWeight_EACL_2024, title = {What the Weight?! A Unified Framework for Zero-Shot Knowledge Composition}, author = {Carolin, Holtermann and Frohmann, Markus and Rekab-saz, Navid and Lauscher, Anne}, booktitle = {Findings of the Association for Computational Linguistics: EACL 2024}, editor = {Yvette Graham, Matthew Purver}, location = {St. Julian’s, Malta}, volume = {18th}, pages = {1138–1157}, month = {March 17-22}, year = {2024} }