Chrome Extension
WeChat Mini Program
Use on ChatGLM

DNA-GPT: Divergent N-Gram Analysis for Training-Free Detection of GPT-Generated Text

ICLR 2024(2024)

University of California Department of Computer Science | NEC-Labs | University of California | UC Santa Barbara

Cited 98|Views177
Abstract
Large language models (LLMs) have notably enhanced the fluency and diversity of machine-generated text. However, this progress also presents a significant challenge in detecting the origin of a given text, and current research on detection methods lags behind the rapid evolution of LLMs. Conventional training-based methods have limitations in flexibility, particularly when adapting to new domains, and they often lack explanatory power. To address this gap, we propose a novel training-free detection strategy called Divergent N-Gram Analysis (DNA-GPT). Given a text, we first truncate it in the middle and then use only the preceding portion as input to the LLMs to regenerate the new remaining parts. By analyzing the differences between the original and new remaining parts through N-gram analysis in black-box or probability divergence in white-box, we can clearly illustrate significant discrepancies between machine-generated and human-written text. We conducted extensive experiments on the most advanced LLMs from OpenAI, including text-davinci-003, GPT-3.5-turbo, and GPT-4, as well as open-source models such as GPT-NeoX-20B and LLaMa-13B. Results show that our zero-shot approach exhibits state-of-the-art performance in distinguishing between human and GPT-generated text on four English and one German dataset, outperforming OpenAI's own classifier, which is trained on millions of text. Additionally, our methods provide reasonable explanations and evidence to support our claim, which is a unique feature of explainable detection. Our method is also robust under the revised text attack and can additionally solve model sourcing.
More
Translated text
Key words
Text Detection,zero-shot,AI detection,GPT
PDF
Bibtex
AI Read Science
AI Summary
AI Summary is the key point extracted automatically understanding the full text of the paper, including the background, methods, results, conclusions, icons and other key content, so that you can get the outline of the paper at a glance.
Example
Background
Key content
Introduction
Methods
Results
Related work
Fund
Key content
  • Pretraining has recently greatly promoted the development of natural language processing (NLP)
  • We show that M6 outperforms the baselines in multimodal downstream tasks, and the large M6 with 10 parameters can reach a better performance
  • We propose a method called M6 that is able to process information of multiple modalities and perform both single-modal and cross-modal understanding and generation
  • The model is scaled to large model with 10 billion parameters with sophisticated deployment, and the 10 -parameter M6-large is the largest pretrained model in Chinese
  • Experimental results show that our proposed M6 outperforms the baseline in a number of downstream tasks concerning both single modality and multiple modalities We will continue the pretraining of extremely large models by increasing data to explore the limit of its performance
Try using models to generate summary,it takes about 60s
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Related Papers
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper

要点】:本文提出了一种名为DNA-GPT的无训练检测策略,通过黑盒或白盒的N-gram分析差异,有效区分人类生成文本和GPT生成文本,并在多个英语和德语数据集上表现优于现有方法,同时提供了可解释的检测证据。

方法】:方法采用DNA-GPT,通过截断文本并使用前部分输入到语言模型中再生新部分,然后通过N-gram分析或概率散度分析原文本与再生文本的差异。

实验】:实验在OpenAI的最先进LLMs上进行,包括text-davinci-003、GPT-3.5-turbo、GPT-4,以及开源模型GPT-NeoX-20B和LLaMa-13B。结果显示,在没有训练的情况下,该零样本方法在四个英语和 one 个德国数据集上表现卓越,效果优于 OpenAI 的训练分类器,并且能抵抗修订文本攻击,同时可用于解决模型来源问题。