Your Finetuned Large Language Model is Already a Powerful Out-of-distribution Detector
arxiv(2024)
摘要
We revisit the likelihood ratio between a pretrained large language model
(LLM) and its finetuned variant as a criterion for out-of-distribution (OOD)
detection. The intuition behind such a criterion is that, the pretrained LLM
has the prior knowledge about OOD data due to its large amount of training
data, and once finetuned with the in-distribution data, the LLM has sufficient
knowledge to distinguish their difference. Leveraging the power of LLMs, we
show that, for the first time, the likelihood ratio can serve as an effective
OOD detector. Moreover, we apply the proposed LLM-based likelihood ratio to
detect OOD questions in question-answering (QA) systems, which can be used to
improve the performance of specialized LLMs for general questions. Given that
likelihood can be easily obtained by the loss functions within contemporary
neural network frameworks, it is straightforward to implement this approach in
practice. Since both the pretrained LLMs and its various finetuned models are
available, our proposed criterion can be effortlessly incorporated for OOD
detection without the need for further training. We conduct comprehensive
evaluation across on multiple settings, including far OOD, near OOD, spam
detection, and QA scenarios, to demonstrate the effectiveness of the method.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要