all search terms
2024 年 10 月 14 日
Hypothesisonly Biases in Large Language ModelElicited Natural Language Inference
title: Hypothesisonly Biases in Large Language ModelElicited Natural Language Inference
publish date:
2024-10-11
authors:
Grace Proebsting et.al.
paper id
2410.08996v1
download
abstracts:
We test whether replacing crowdsource workers with LLMs to write Natural Language Inference (NLI) hypotheses similarly results in annotation artifacts. We recreate a portion of the Stanford NLI corpus using GPT-4, Llama-2 and Mistral 7b, and train hypothesis-only classifiers to determine whether LLM-elicited hypotheses contain annotation artifacts. On our LLM-elicited NLI datasets, BERT-based hypothesis-only classifiers achieve between 86-96% accuracy, indicating these datasets contain hypothesis-only artifacts. We also find frequent “give-aways” in LLM-generated hypotheses, e.g. the phrase “swimming in a pool” appears in more than 10,000 contradictions generated by GPT-4. Our analysis provides empirical evidence that well-attested biases in NLI can persist in LLM-generated data.
QA:
coming soon
编辑整理: wanghaisheng 更新日期:2024 年 10 月 14 日