Abstract
The rapid adoption of AI in academic writing, particularly with tools like ChatGPT, has raised significant concerns regarding the accuracy of generated content. This study explores the phenomenon of “hallucitation” in scientific writing, where AI models fabricate citations, analyzing responses from ChatGPT versions 3.5 and 4o in the context of librarianship. Through an experimental design, scientific content with citations was generated and systematically verified using Google Scholar and the publisher’s website. The findings reveal a disturbingly high frequency of false or non-existent citations—42.9% in ChatGPT-3.5 and 51.8% in ChatGPT-4o. Despite slight improvements in citation accuracy from version 3.5 to 4o, with accuracy rates of 3.92% and 6.35%, respectively, both versions exhibit significant limitations. Notably, ChatGPT 3.5 frequently generated completely fabricated sources, while ChatGPT-4o introduced subtle errors, such as mismatched journals. The study indicates no significant difference in accuracy between the two versions, underscoring the persistent risks associated with AI-generated citations. These findings highlight the urgent need for rigorous verification of AI-generated content to safeguard the integrity of scholarly work.
| Original language | English |
|---|---|
| Pages (from-to) | 62-92 |
| Number of pages | 31 |
| Journal | Journal of Web Librarianship |
| Volume | 19 |
| Issue number | 1 |
| DOIs | |
| Publication status | Published - 2025 |
Keywords
- AI-generated citations
- ChatGPT
- Hallucination
- academic integrity
- librarianship
ASJC Scopus subject areas
- Computer Science Applications
- Library and Information Sciences
Fingerprint
Dive into the research topics of 'Hallucitation in Scientific Writing: Exploring Evidence from ChatGPT Versions 3.5 and 4o in Responses to Selected Questions in Librarianship'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver