One of Caveon’s professional services is Secret Shopping, conducted at the request and with the cooperation of testing programs. During a Secret Shopping engagement, a trained evaluator books a real exam slot and behaves like an ordinary candidate while probing for gaps in live security. Some log in from a bedroom laptop, others check in at busy test centers.
Their mission is simple: behave like an ordinary candidate while attempting the kind of low-tech violations that any determined test-taker could try. The goal is not to embarrass proctors––most are diligent, polite, and doing their best––but to test whether the security measures that programs rely on actually work.
The data and stories below are typical of what we observe in these experiments day after day.
These observations, and the failings they expose, frustrate us, because they show how often well-meaning programs depend on safeguards that cannot deliver the security they promise.
Think of ‘secret shopping’ as the assessment world’s version of an undercover health and safety inspector or a penetration test in cybersecurity. An inspector might try to purchase alcohol without showing their ID to evaluate whether everyone is being carded. In IT, a red-team specialist might see if a mock virus can make it through a company’s firewalls. These measures help vet the quality of security measures to ensure that they can effectively catch the threats they’re designed to stop.
In testing, ‘secret shoppers’ are brought in to evaluate the efficacy of proctors and the exam administration protocols they have in place. These trained agents enter an exam under genuine credentials, undergo the standard check-in process, and then attempt to perform certain simple violations throughout the exam. For example, they use their phones, sneak in earbuds or smart watches, open notes, dig in their bags, or leave prohibited items on their desks, etc. Throughout the exam, they log whether or not these attempts are caught by the exam proctor, and if so, which remedial actions are taken.
Before diving into two representative stories, it helps to see the wider picture. At the 2024 National Council on Measurement in Education, Caveon shared aggregate results from our past secret shop missions. Across engagements, one pattern dominates: proctors—whether in person or online—miss the majority of visible cheating attempts, and when they do notice, the session often continues after a warning to the test-taker.
The numbers confirm that these incidents are not outliers but the norm.
These results underline how easily everyday tactics slip past proctors.
Table 1: Testing infractions attempted by one secret shopper during a remote proctored exam.
This secret shopping engagement detailed below took place in a testing center with an in-person proctor. The “undercover” test-taker was screened and then entered the testing environment, just like any other examinee.
In this instance, the test-taker was able to hide notes in the bathroom, take an unauthorized bathroom break to read the hidden notes, write down the exam questions they had seen and then return to the testing environment, hide a cell phone in their waistband, and record more than 25 minutes of the testing session.
Table 2: Testing infractions attempted by one secret shopper during an in-person proctored exam.
Even when proctors are well trained and diligent, six recurring factors open gaps large enough for routine cheating to slip through.
Because these fault lines play out in almost every session, Caveon’s secret shoppers routinely pass nine out of ten visible cheating attempts. Had they been intent on real fraud, they could have walked away with high scores and high-value item banks, leaving the program exposed to costly score invalidations and reputational damage once stolen content surfaced online.
Caveon shares these narratives not to blame or shame proctors, most of whom work hard within the limits of their tools, but rather to show why those limits matter. Programs that depend on proctoring alone may feel secure, yet the evidence says otherwise. Layered defenses—secure exam design, real-time analytics, continuous forensics—offer a clearer path.



