Feedback boosts human ability to spot AI-generated texts

Humans struggle to detect AI-generated texts but can improve with feedback, a new study finds.

Why it matters

  • Humans are initially poor at detecting AI-generated texts.
  • Feedback improves detection accuracy by 10 percentage points.
  • Overconfidence without feedback leads to more errors.

By the numbers

  • 254 Czech native speakers participated.
  • No-feedback group accuracy: 55.4%.
  • Feedback group accuracy: 65.1%.
  • Feedback group improved over time (p < 0.001).

The big picture

  • Humans have misconceptions about AI-generated text features.
  • Feedback helps correct these misconceptions and improves accuracy.
  • Detection skills can be learned and improved with practice.

What they're saying

  • Participants initially assumed AI-generated texts were more readable.
  • Without feedback, overconfidence led to more errors.
  • Commenters noted language-specific quirks and the potential obsolescence of detection methods.

Caveats

  • Study conducted in Czech; results may vary by language.
  • No significant effect of AI usage frequency or attitudes on detection ability.

What’s next

  • Potential for educational tools to help detect AI-generated content.
  • Continued research on AI detection methods as AI models evolve.