Virginia Tech research warns that unproctored online testing has become entirely vulnerable to cheating thanks to advanced large language models such as ChatGPT. The study, led by industrial organizational psychologist Louis Hickman, highlights that developments in reasoning large language models—including OpenAI´s ´o1´ model—have rendered traditional unsupervised digital assessments obsolete, as anyone can now copy-paste test questions into these models to receive high-performing answers in seconds.
Unproctored assessments have long enabled employers and educators to evaluate wide groups of candidates and students efficiently. However, the research shows that new reasoning models, trained with reinforcement learning and capable of internal monologue for self-improvement, have moved beyond earlier limitations—particularly on quantitative ability tests. The latest generation of models achieves strong results even on complex assessments, including personality, situational judgment, verbal ability, and numerical reasoning, effectively undermining the integrity and predictive value of such tests. Industry surveys indicate that about one-third of job applicants and most students are leveraging large language models for critical assessments and coursework.
With the validity of unproctored testing declining, experts now recommend essential overhauls in test administration and design. Proposed solutions include reverting to supervised proctoring, integrating large language models into the assessment process, imposing strict time constraints, employing software to detect artificial intelligence usage, requesting test-takers to verbalize thought processes, and analyzing digital traces for authenticity. While each approach has trade-offs, the research makes clear that employers and educators can no longer rely on legacy testing protocols and must adapt to the realities posed by generative artificial intelligence.