fromNature
1 week agoPsychology
Evaluating large language models for accuracy incentivizes hallucinations - Nature
Large language models produce confident falsehoods due to evaluation metrics that reward guessing over uncertainty, necessitating new evaluation methods.