DecLimSup: Decoding with Limited Teacher Supervision Requires Understanding When to Trust the Teacher
You can run the code by referring to the script.sh file
Also our code supports Gaudi HPUs. You can enable it by adding the --use_hpu.
You can run the code using the following command:
python inference.pyBelow is a description of the key arguments:
- Description: Specifies the benchmark to evaluate the model on.
- Supported Benchmarks:
gsm8k,strategyqa,multiarith,math,arc_c,arc_e,svamp
- Description: Determines how many tokens will receive knowledge from the teacher model.
-
Description: Enables running multiple alpha values in a single execution.
DecLimSup is our work that empirically analyzes contrastive decoding settings in a limited supervision scenario of teacher LLM. We find that it is essential to adaptively overtrust or disregard the LLM prediction based on the confidence of the small-scale LLM. Our experiments on a wide range of models and datasets demonstrate that our method consistently improves over conventional decoding strategies.
If you use this code, please cite the following paper:
@inproceedings{ok-etal-2024-decoding,
title = "Decoding with Limited Teacher Supervision Requires Understanding When to Trust the Teacher",
author = "Ok, Hyunjong and
Ryu, Jegwang and
Lee, Jaeho",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2024.emnlp-main.693",
pages = "12460--12476",
}