Asking chatbots for short answers can increase hallucinations, study finds
: tors maj 08, 2025 12:05 pm
Turns out, telling an AI chatbot to be concise could make it hallucinate more than it otherwise would have. That’s according to a new study from Giskard, a Paris-based AI testing company developing a holistic benchmark for AI models. In a blog post detailing their findings, researchers at Giskard say prompts for shorter answers to […]
Source: https://techcrunch.com/2025/05/08/askin ... udy-finds/
Source: https://techcrunch.com/2025/05/08/askin ... udy-finds/