Gap Between Fluency and Reasoning
ChatGPT-5 Earns a 'D' Grade for Scientific Accuracy
Washington State University researchers found the AI model struggles with scientific consistency and reasoning.

A digital tablet on a wooden desk showing a document with a red 'D' grade circled, next to a holographic representation of a neural network.
Photo: Avantgarde News
Washington State University researchers graded ChatGPT-5 a "D" for its scientific accuracy and consistency [1]. The study tested the AI against more than 700 scientific hypotheses to evaluate its performance [1]. The results revealed a significant gap between the model's linguistic fluency and its conceptual reasoning [1]. While the AI generates articulate text, it frequently fails to identify false statements correctly [1]. According to the study, these inconsistencies suggest that current AI models lack a deep understanding of scientific logic [1]. This research highlights the need for caution when using AI for verifying complex information [1].
Editorial notes
Transparency note
Drafted with LLM; human-edited
- AI assisted
- Yes
- Human review
- Yes
- Last updated
Risk assessment
The risk level is set to high because the story relies on a single source domain (WSU Insider), which fails the recommendation for three independent domains.
Sources
Related stories
View allTopics
About the author
Avantgarde News Desk covers gap between fluency and reasoning and editorial analysis for Avantgarde News.


