Symbol tuning improves in-context learning in language models

J Wei, L Hou, A Lampinen, X Chen… - Proceedings of the …, 2023 - aclanthology.org
Proceedings of the 2023 Conference on Empirical Methods in Natural …, 2023aclanthology.org
We present symbol tuning-finetuning language models on in-context input-label pairs where
natural language labels (eg,“positive/negative sentiment”) are replaced with arbitrary
symbols (eg,“foo/bar”). Symbol tuning leverages the intuition that when a model cannot use
instructions or natural language labels to figure out a task, it must instead do so by learning
the input-label mappings. We experiment with symbol tuning across PaLM models up to
540B parameters and observe benefits across various settings. First, symbol tuning boosts …
Abstract
We present symbol tuning-finetuning language models on in-context input-label pairs where natural language labels (eg,“positive/negative sentiment”) are replaced with arbitrary symbols (eg,“foo/bar”). Symbol tuning leverages the intuition that when a model cannot use instructions or natural language labels to figure out a task, it must instead do so by learning the input-label mappings. We experiment with symbol tuning across PaLM models up to 540B parameters and observe benefits across various settings. First, symbol tuning boosts performance on unseen in-context learning tasks and is much more robust to underspecified prompts, such as those without instructions or without natural language labels. Second, symbol-tuned models are much stronger at algorithmic reasoning tasks, with up to 18.2% better performance on the List Functions benchmark and up to 15.3% better performance on the Simple Turing Concepts benchmark. Finally, symbol-tuned models show large improvements in following flipped-labels presented in-context, meaning that they are more capable of using in-context information to override prior knowledge.
aclanthology.org
Showing the best result for this search. See all results