Next Article in Journal
Creative and Critical Thinking and Modelling: Confluences and Implications for Science Teaching
Previous Article in Journal
Emotional Intelligence and the Big Five as Predictors of Students’ Performance in Collaborative Problem Solving
Previous Article in Special Issue
Toward a Consensus Model of Cognitive–Reading Achievement Relations Using Meta-Structural Equation Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effects of Proctoring on Online Intelligence Measurement: A Literature Overview and an Empirical Study

Department of Psychology, Trier University, D-54286 Trier, Germany
*
Author to whom correspondence should be addressed.
J. Intell. 2025, 13(9), 110; https://doi.org/10.3390/jintelligence13090110
Submission received: 31 July 2025 / Revised: 20 August 2025 / Accepted: 28 August 2025 / Published: 30 August 2025
(This article belongs to the Special Issue Intelligence Testing and Assessment)

Abstract

Remote intelligence testing has multiple advantages, but cheating is possible without proper supervision. Proctoring aims to address this shortcoming, yet prior research on its effects has primarily investigated reasoning tasks, in which cheating is generally difficult. This study provides an overview of recent research on the effects of proctoring and on studies in intelligence test settings. Moreover, we conducted an empirical study testing the effects of webcam-based proctoring with a multidimensional intelligence test measuring reasoning, short-term memory, processing speed, and divergent thinking. The study was conducted in a low-stakes context, with participants receiving a fixed payment regardless of performance. Participants completed the test under proctored (n = 74, webcam consent), unproctored random (n = 75, webcam consent), or unproctored chosen (n = 77, no webcam consent) conditions. Scalar measurement invariance was observed for reasoning, processing speed, and divergent thinking, but not for memory. Proctoring had no significant main effect on test performance but showed a significant interaction with test type. Proctored participants outperformed the unproctored chosen group significantly in divergent thinking and scored descriptively higher in reasoning and processing speed, but slightly lower in memory. Observable cheating under proctored conditions was rare (4%), mostly involving note-taking or photographing the screen. We conclude that proctoring is crucial for easily cheatable tasks, such as memory tasks, but currently less critical for complex cognitive tasks.

1. Introduction

Intelligence testing is one of the cornerstones of psychological assessment. It is widely used to support placement and intervention decisions in a variety of applied settings (Nettelbeck and Wilson 2005; Ones et al. 2017; Taylor et al. 2008). Furthermore, intelligence testing is employed in research spanning numerous academic disciplines, including psychology, education, sociology, economics, and medicine (Fletcher and Hattie 2011; Mackintosh 1998). Intelligence tests assess maximum performance, meaning they aim to capture the highest level of cognitive performance the test taker can achieve. Therefore, underperformance due to factors such as low motivation or adverse circumstances during testing, as well as artificial increases in performance due to drug use or cheating, pose a threat to their validity. In intelligence testing, cheating refers to intentional test-taker behavior that violates standardized conditions and undermines test fairness and validity. In online testing, this can include behaviors such as the use of unauthorized aids like internet searches, dictionaries, or calculators (Steger et al. 2018). Other forms include collaborating with others during a test, having another person complete the test, gaining prior access to test questions or materials, and using physical resources such as books, cameras, or handwritten notes during an online assessment (Öncül 2021).
To minimize adverse circumstances and cheating, intelligence tests are usually conducted in controlled environments under the supervision of a trained test administrator. However, the rise in computer-based assessment has made it possible to instead test individuals using their computers in their own homes (i.e., remote testing). This is an attractive option due to lower costs and reduced participation burden, especially in research contexts where large samples for cognitive testing are often needed, but difficult to obtain. At the same time, remote testing hinders environmental control and supervision, particularly raising concerns about cheating. Various methods have been developed to prevent or detect cheating, such as honor codes (e.g., Karim et al. 2014; Tatum and Schwartz 2017), specialized software environments (e.g., Mohammed and Ali 2022), and statistical cheating detection (e.g., Kingston and Clark 2014). One of the most common methods is proctoring, where the behavior and environment of test subjects is observed directly during testing, including webcam recording of the test taker, screen capture, or keystroke logging (Steger et al. 2018). Implementing a proctoring solution in remote testing requires participants’ willingness to be monitored via a webcam feed and increases costs and participation burden compared to unproctored assessments, which partially negates the advantages of remote testing. Therefore, understanding the effects, effectiveness, and necessity of proctoring in remote testing is of interest to researchers and practitioners who must decide on the mode of testing for their study or diagnostic application.
Previous studies examining the effects of proctoring on intelligence testing have predominantly focused on reasoning tasks, which are not prone to cheating. However, a wide range of cognitive tasks exist that are relevant both for cognitive research and for comprehensive assessments of cognitive abilities in applied contexts. In this study we therefore investigated the effects of proctoring on four distinct abilities: reasoning, short-term memory, processing speed, and divergent thinking. Moreover, previous research usually compared test results between proctored and unproctored conditions. A precondition for doing so is that the test measures cognitive abilities equivalently across conditions (i.e., measurement invariance). There are very few findings for measurement invariance in proctoring studies. In our study, we therefore tested the measurement invariance of the intelligence test across different conditions to contribute to closing this gap.

1.1. Proctoring

Proctoring is the practice of supervising test takers during online exams or cognitive testing to prevent or detect dishonest behavior, i.e., cheating. One proctoring method is live video observation, in which a test administrator watches the test taker and their environment via a webcam feed. In some cases, the feed is recorded and watched later. A second method is artificial intelligence (AI) based proctoring, where the video feed is observed by a trained AI system that flags potential instances of cheating (Langenfeld 2020). In all cases, the aim is to deter cheating by making test takers aware that they are being observed and to detect any cheating.

1.1.1. Reviews and Meta-Analyses on the Effects of Proctoring

Steger et al. (2018) conducted a meta-analysis of the impact of proctoring on ability assessments, including 49 studies published between 2001 and 2017. The analysis covered studies of low- and high-stakes testing situations involving paper-and-pencil or computerized tests. Most of the studies were conducted in educational or research contexts, while the remainder came from industrial and organizational contexts. The results showed that performance on unproctored assessments was superior to performance on proctored assessments, with a modest standardized mean difference effect size of d = 0.20. A key limitation, however, is that the original studies did not report measurement invariance of the tests across conditions. As a result, measurement invariance could not be investigated in the meta-analysis. Note that measurement invariance is not only a prerequisite for comparing means between groups (Cheung and Rensvold 2002; Chen 2007) but also allows researchers to examine which properties of individual indicators differ between proctoring conditions. For example, if some test items are easier to cheat on due to higher searchability, these items would be expected to be flagged in measurement invariance testing by showing higher intercepts on the intelligence factor in the non-proctored condition. This would indicate that such items, relative to all other items, are easier to solve in the non-proctored condition than in the proctored condition. Therefore, investigating measurement invariance is essential when comparing proctored and non-proctored conditions.
Moderator analyses identified online searchability as the only significant moderator, revealing larger mean score differences for tasks that could be easily solved through internet searches. This finding led Steger et al. (2018) to recommend unproctored settings only for tasks that cannot easily be solved using online searches. No significant effects were found for the other moderators: stakes, countermeasures to cheating, and test mode (i.e., paper–pencil vs. computerized). The lack of significant differences in the results of high- and low-stakes conditions indicated that cheating may occur in both conditions when unproctored. In addition to cheating, the authors proposed reduced test anxiety in unproctored participants as an explanation for score differences between proctored and unproctored conditions.
Following the shift towards online educational settings prompted by the COVID-19 pandemic, Öncül (2021) published a systematic review comparing proctored and unproctored exams. The review incorporated 15 quantitative, qualitative, and meta-analytic articles published between 2004 and 2020. Most of these studies examined college students in the USA, except for Domínguez et al. (2019) and Steger et al.’s (2018) meta-analysis. Öncül found mixed results regarding differences in test scores between proctored and unproctored exams. Eight studies reported a significant difference, with unproctored exam scores being higher in seven of them, while the remaining studies found no difference. The potential reasons for these discrepancies in results were largely unexplored.

1.1.2. Overview of Recent Studies

In light of rapid advancements in the field in recent years, we searched for empirical studies published since Steger et al.’s meta-analysis in 2018 Steger et al. (2018). Table 1 presents 15 studies that were identified through an online search. Only four of these studies were also included in Öncül (2021). Most of the recent studies focused on academic testing and were conducted in high-stakes academic settings, involving university students as participants. The majority examined samples from the US, with fewer studies from China, Indonesia, Mexico, Norway, and Spain. The studies varied in terms of the assessment location (i.e., on-site vs. remote) and the proctoring solutions used (i.e., webcam recording vs. automated proctoring software vs. on-site proctoring). In addition to reporting differences in test performance between proctored and unproctored conditions (14 studies), some studies also reported evidence of cheating (5 studies), differences in test duration (4 studies), and measurement invariance findings (1 study).
Of the 14 studies that examined differences in test performance, the majority (9; 64%) reported poorer test performance in proctored than unproctored conditions (Alessio et al. 2018; Baso 2022; Chen et al. 2020; Daffin and Jones 2018; Dendir and Maxwell 2020; Domínguez et al. 2019; Oeding et al. 2024a; Reisenwitz 2020; Rodríguez-Villalobos et al. 2023). This difference was sometimes more pronounced for more difficult questions (Alessio et al. 2018; Chen et al. 2020) or when the test situation was perceived as high-stakes (Alessio et al. 2018). Two studies found no difference in test performance between the two conditions (Chan and Ahn 2023; Norrøne and Nordmo 2025), while the remaining three studies produced mixed results (Feinman 2018; Oeding et al. 2024b; Zhang et al. 2024) with two reporting better performance in the proctored condition for some of their samples or tests (Feinman 2018; Zhang et al. 2024).
Although better performance in unproctored tests was sometimes considered to be an indication of cheating, only five studies have explicitly investigated cheating behaviors, all in high-stakes settings (Chan and Ahn 2023; Chen et al. 2020; Dendir and Maxwell 2020; Norrøne and Nordmo 2025; Zhao et al. 2024). These studies generally reported low levels of cheating. However, in a longitudinal study, Chen et al. (2020) observed that the score advantage of the unproctored condition, the number of cheaters, and the benefit of cheating increased over the course of the semester, which the authors explained by an increase in cheating proficiency. Focusing exclusively on cheating, Zhao et al. (2024) surprisingly reported a greater frequency of cheating in the proctored condition than in the unproctored condition. The authors recommended adding academic integrity reminders before the exam, which significantly reduced cheating in both the proctored and unproctored conditions.
In terms of test duration, participants were generally found to perform faster in proctored conditions than in unproctored ones (Alessio et al. 2018; Daffin and Jones 2018; Zhang et al. 2024). Daffin and Jones (2018) found that participants in the proctored conditions took only half as long to complete the exam. Conversely, Domínguez et al. (2019) reported no difference between the two settings but found that participants spent more time on more difficult tasks in the unproctored condition compared to the proctored condition.
Finally, only Norrøne and Nordmo (2025) examined measurement invariance in their analyses, reporting strict measurement invariance over conditions. Of note, the lack of measurement invariance testing limits the interpretability of the reported differences in performance levels in most of the studies reported here. Seven years after Steger et al. (2018) raised the issue in their meta-analysis, this frequent omission remains a prevailing limitation of proctoring research.

1.1.3. Proctoring in Intelligence Testing

Most studies included in the meta-analysis by Steger et al. (2018), the review by Öncül (2021), and our own review of recent studies focused on academic testing. The generalizability of these results to intelligence testing, particularly in low-stakes research settings, has rarely been examined, with a few notable exceptions. Table 2 presents nine studies that investigated the effects of proctoring in intelligence test settings. Most of them used reasoning tests, assessing figural content such as Raven’s matrices, verbal content such as word similarities, or numeric content such as number reasoning. However, other tests such as knowledge questions, mental rotation, and creativity tests were also applied. The investigated samples were diverse, as were the countries of sample origin (3 × USA, 2 × UK, Germany, Norway, Netherlands, Singapore). Five studies examined low-stakes testing situations, while four focused on high-stakes situations involving job applicants. Most studies compared on-site proctoring with online, unproctored testing conditions. In addition to reporting differences in average test performance between proctored and unproctored conditions (9 studies), some studies also reported evidence of cheating (5 studies), mean differences in test duration (1 study), and measurement invariance findings (2 studies).
Regarding average test performance, most studies reported no significant differences between proctored and unproctored conditions (5 studies: Ihme et al. 2009; Lievens and Burke 2011; Norrøne and Nordmo 2025; Schakel 2012; Templer and Lange 2008). Other studies found mixed results, with unproctored participants performing better on certain tests or indicators but showing no differences or even worse performance on others (3 studies: Karim et al. 2014; Williamson et al. 2017; Wright et al. 2014). Coyne et al. (2005) reported better results in the proctored condition. Karim et al. (2014) considered the searchability of test content (e.g., knowledge questions versus reasoning tests) as a moderator and found that the advantage for the unproctored condition occurred only in the high searchability condition. This finding may partially explain why most studies found no difference regardless of test stakes, as most studies used reasoning tests with low online searchability.
Consistent with the lack of performance benefits in unproctored settings, most studies revealed limited or no evidence of cheating (Lievens and Burke 2011; Norrøne and Nordmo 2025; Wright et al. 2014). Schakel (2012) observed moderate indication of cheating. Karim et al. (2014) reported higher levels of cheating in the unproctored condition than in the proctored condition using a highly searchable test.
Only one study (Wright et al. 2014) examined test duration and found mixed results. In one sample, unproctored participants had shorter test durations, whereas in another sample, they took longer to complete the test compared to proctored participants. Two studies (Norrøne and Nordmo 2025; Wright et al. 2014) considered measurement invariance. Norrøne and Nordmo (2025) observed the strict measurement invariance for their test. Wright et al. (2014) found that 8 out of 30 items showed differential item functioning. Notably, some items were slightly easier in the proctored condition, while others were easier in the unproctored group, which argues against systematic cheating across the entire scale.

1.1.4. Summary and Open Questions

Overall, the available evidence suggests small yet significant performance advantages in unproctored conditions. However, this advantage only seems to exist for tests where answers can easily be found online, such as academic exams or knowledge tests. The advantage was not found for most intelligence tests, particularly reasoning tests, which were by far the most prevalent test format. Cheating levels were found to be low in all conditions. Test duration and measurement invariance were rarely considered in the studies.
Although the available findings provide valuable guidance for researchers and practitioners, some questions remain. First, previous research has only distinguished between searchable and non-searchable test formats. However, there are other ways to cheat besides looking up answers to knowledge questions online. For instance, memory tests allow cheating through note-taking or taking photos of the screen. Other tasks, such as verbal reasoning or divergent thinking tasks, can be solved with the help of large language models (LLMs), like ChatGPT or Claude. Finally, some tasks, such as for example, processing speed tasks, which are simple and scale through response speed, or figural matrix tasks, which require complex inductive reasoning that cannot be easily conveyed, do not lend themselves to cheating because any attempt incurs time costs without improving performance. Thus, when it comes to distinguishing between tasks, the concept of searchability falls short. A more general classification based on cheatability can cover additional cases.
Second, most findings on average performance differences between proctored and unproctored conditions do not rely on prior tests of measurement invariance. This is problematic because violations of scalar measurement invariance can bias mean-level comparisons (Borsboom 2006; Schmitt and Kuljanin 2008). Measurement invariance examines whether the same set of indicators measures an identical latent construct equivalently across different groups or conditions. Different levels of measurement invariance can either be established or violated (Cheung and Rensvold 2002; Little 1997). The lowest level, configural invariance, implies that the same measurement model fits both conditions. Metric invariance assumes equal factor loadings across conditions, allowing comparisons of latent correlations and regressions between groups in different conditions. Scalar invariance tests whether intercepts of indicators are identical across conditions, allowing comparisons of latent means. Non-invariant intercepts across conditions suggest that specific indicators differ in difficulty relative to the overall scale. For example, in unproctored conditions, certain indicators might become easier due to increased opportunities for cheating (e.g., higher online searchability compared to other items). If scalar measurement invariance is not established, observed mean differences between groups remain ambiguous, as it is unclear whether differences reflect true differences in the latent construct or differences arising from indicators measuring distinct constructs across conditions. Therefore, previous findings regarding performance differences between proctored and unproctored conditions should be interpreted cautiously, given the possibility that identical performance indicators might not measure the same construct in both conditions (Steger et al. 2018).

1.2. The Present Study

Previous studies examining the effects of proctoring on intelligence testing have predominantly focused on reasoning tasks. However, a wide range of other cognitive tasks exist that are relevant both for cognitive research and for comprehensive assessments of cognitive abilities in applied contexts. Cognitive tasks such as memory or processing speed tasks differ in the ease with which participants can engage in cheating behavior during remote testing and in the types of cheating strategies that may be employed to solve them (e.g., note-taking vs. online search). To consider differences in cheatability, we investigated the effects of proctoring on four distinct intelligence abilities (i.e., reasoning, short-term memory, processing speed, divergent thinking). For reasoning and divergent thinking, we expected low cheatability, as cheating in these tasks would likely require support from AI tools or assistance from other individuals. For short-term memory, we expected high cheatability due to the potential for note-taking, either by writing down or photographing items with a mobile phone. For processing speed, we expected no cheatability, as there are no known cheating strategies that would meaningfully improve test performance.
We used a typical research scenario, involving online remote testing under low-stakes conditions. Specifically, we compared three experimental conditions: a proctored condition with webcam video recording, an unproctored random condition (UP random; participants were generally willing to be proctored with webcam video recording but were randomly assigned to not being proctored), and an unproctored chosen condition (UP chosen; participants explicitly declined proctoring with webcam video recording and were therefore not proctored). Including these three conditions was necessary for two reasons: First, it allowed us to experimentally assign participants to either a proctored or unproctored setting. Second, it enabled us to investigate whether individuals who explicitly chose not to be proctored systematically differed in performance from those who agreed to proctoring. In total, we addressed four research questions.
First, for each of the four ability test scales we examined three levels of measurement invariance (configural, metric, and scalar) across conditions. To the best of our knowledge, only two studies have investigated measurement invariance in intelligence testing across proctoring conditions before (Norrøne and Nordmo 2025; Wright et al. 2014). Norrøne and Nordmo found support for the highest level of invariance, but this was examined only for assessments of reasoning ability, where cheatability is low. Wright et al. found that certain items of a speeded application test showed different psychometric properties across proctoring conditions. Therefore, we made no specific a priori assumptions regarding measurement invariance for the current study but assessed measurement invariance as an open research question.
Second, we investigated mean differences in test performance across conditions and whether the effects of proctoring varied by ability test (i.e., the interaction between condition and test type). For short-term memory tests, we hypothesized that mean scores would be lower in the proctored condition compared to both the UP random and UP chosen conditions, due to the possibility of cheating through note-taking in unproctored settings. For reasoning and divergent thinking tests, we made no specific a priori predictions, as cheating is possible but considered rather difficult. For processing speed, we expected no differences between proctoring conditions, as cheatability for this type of task should not be possible.
Third, we examined test duration across proctoring conditions as an open research question. Although cheating could theoretically impact test duration, existing empirical evidence regarding this relationship is inconclusive. It should be noted that only the total test duration was available in our study, so we were unable to distinguish test times for the individual ability tests.
Fourth, we explored how many participants demonstrated cheating behavior in the proctored condition with video monitoring. According to previous research (e.g., Lievens and Burke 2011; Norrøne and Nordmo 2025; Wright et al. 2014), cheating is relatively rare. However, little is known about the prevalence of cheating in low-stakes remote intelligence testing with webcam proctoring.

2. Methods

2.1. Participants and Procedure

Data collection for the present study was conducted online between late April and early June 2025, using the Prolific platform as part of a larger data collection aimed at piloting the new digital intelligence test BIS+ (Berlin Structure-of-Intelligence Test Plus; developed by the authors of this paper). Prolific is an online platform through which participants can take part in scientific studies or commercial surveys for monetary compensation. All participants were recruited from Germany and self-reported fluency in German was a pre-requirement for study participation. The data collection was approved by the ethics committee of Trier University (protocol numbers 46/2025).
Overall, 266 test takers started the study; 40 of them dropped out and did not complete the tasks (n = 13 in the proctored condition; n = 9 in UP random condition; n = 18 in UP chosen condition). Thus, valid data from a total of 226 participants were available for analysis (age: M = 29.61, SD = 6.41, Min = 20, Max = 44; gender: 75 female, 147 male, 4 non-binary).
Participants were able to choose the day and time of their participation independently. The study description clearly communicated that participation required approximately 155 min of uninterrupted work in a quiet environment with full concentration. Participants received compensation of £25.84 (£10 per hour). Note that this fixed-amount payment structure, without additional incentives based on performance, may encourage participants to complete the study quickly and “well enough” to be approved, rather than to maximize accuracy. Participants were informed that attentive and earnest participation would help ensure the quality and validity of the test. Additionally, participants were offered individual feedback on their results upon request after completing the study. Participants were also informed that compensation would only be provided after the researchers verified serious participation. Individuals who terminated their participation prematurely did not receive compensation.
Participants who consented to webcam video recording on Prolific were randomly assigned to either (1) the proctored condition or (2) the UP random condition. In the proctored condition, participants were informed prior to testing that they would be recorded during the testing, were indeed recorded during the test, and the videos were subsequently reviewed by our research team. Additional proctoring methods such as screen capture or keystroke logging were not implemented due to limitations in technical feasibility within the Pavlovia system and their potential to reduce participant acceptance. In the UP random condition, participants were informed prior to testing that they would not be recorded, and no video recording took place. Finally, the third condition included participants who indicated on Prolific that they generally did not consent to video recording; these individuals were not recorded (UP chosen). The distribution across conditions was as follows: proctored condition (n = 74), UP random condition (n = 75), and UP chosen condition (n = 77).
Apart from recording or not recording the participants via webcam, the conditions did not differ. Participants first completed a short questionnaire assessing their self-estimated intelligence across various intelligence facets. Subsequently, participants completed the full BIS+ test, which was administered in two parts separated by a short break. Finally, participants completed a second questionnaire that assessed data on demographics, personality traits, and motivational variables.

2.2. Materials

2.2.1. Intelligence Test

Intelligence was assessed using the BIS+, a digital, revised, and shortened adaptation of the BIS-HB (Jäger et al. 2006) that is currently under development. It measures intelligence based on the Berlin Intelligence Structure Model (BIS; Jäger 1984; see Figure 1). The BIS conceptualizes intelligence along four operative abilities, reasoning (9 tasks in the present study), short-term memory (9 tasks), processing speed (9 tasks), and divergent thinking (6 tasks), fully crossed with three content domains: verbal (12 tasks), numeric (12 tasks), and figural (9 tasks). Thus, each process is assessed with material from every content domain, and each domain is assessed across all operative abilities, which are described in Table 3. In this study, we focus solely on the four operative abilities. To estimate reliability, we constructed parcels that each included one figural, one verbal, and one numerical task from each operative ability. The resulting Cronbach’s alpha/Omega values were α = 0.86/ω = 0.86 for reasoning, α = 0.85/ω = 0.85 for short-term memory, α = 0.93/ω = 0.93 for processing speed, and α = 0.71/ω = 0.71 for divergent thinking.

2.2.2. Descriptive Information

Gender was assessed by asking participants to indicate their gender, with response options Female, Male, and Non-Binary. Current employment or student status was measured by asking participants to select the option that best described their situation: School student, Vocational training, University student, Full-time employed, Part-time employed, Self-employed, Military/Civil/Social Service, Unemployed, Incapable of working, Homemaker, or Refugee. Educational degree was assessed by asking participants to indicate the highest degree they had obtained. Response options were: No school degree, Lower secondary school, Intermediate school degree, High school diploma (Abitur), University degree, Doctorate (PhD), and Other degree. Parental education was measured using the same response categories as the participant’s own education. Primary language was assessed with an open-ended question (“Please indicate your primary language”), and answers were subsequently categorized for analysis. Finally, urbanity of residence was assessed by asking participants to indicate the type of area in which they live: Village (<3000 inhabitants), Small town (≥3000; <15,000 inhabitants), Town (≥15,000; <100,000 inhabitants), Large city (≥100,000; <1,000,000 inhabitants), and Metropolis (≥1,000,000 inhabitants).

2.3. Data Analyses

Analysis code and dataset are available online on: https://osf.io/yznct/?view_only=66c66d3627104459926aaf4afa265c74 (Accessed on 17 August 2025). The data analyses were conducted with R 4.4.1. (R Core Team 2024) using the following R packages: apaTables 2.0.8 (Stanley 2021), car 3.1.2 (Fox et al. 2023), dplyr 1.1.4 (Wickham et al. 2014), flextable 0.9.6 (Gohel and Skintzos 2017), ggplot2 3.5.1 (Wickham 2024), lavaan 0.6.18 (Rosseel 2012), psych 2.4.3 (Revelle 2023), purr 1.0.2 (Henry and Wickham 2023), readr 2.1.5 (Wickham et al. 2023), and officer 0.6.6 (Gohel 2023).

2.3.1. Preliminary Analyses

In preliminary analyses, we tested whether the proctoring conditions differed with respect to demographic variables. This step was necessary to ensure that any observed differences in intelligence tests were not simply due to systematic differences in the composition of the groups (e.g., notably younger participants or participants with substantially lower education in one condition). Age was analyzed using one-way ANOVAs with condition as the predictor. Categorical variables such as gender, current employment or student status, highest education, parental education, and primary language were tested using χ2 tests. For each variable, all categories that accounted for less than 5% of cases in the total sample (i.e., less than 12 participants) were combined into an “Other” category to ensure a robust estimation of the χ2-statistic. If the resulting “Other” category accounted for less than 5% of the total sample, it was excluded from the χ2-analyses. For example, in the case of gender, the non-binary category was excluded from the χ2-tests, as it represented only 1.77% of all cases (i.e., 4 out of 226 participants).

2.3.2. Reviewing Proctoring Records

Similarly to Steger et al.’s (2018) recommendation to control post hoc for cheating behavior, we reviewed the available video recordings for evidence of participant cheating to ensure the validity of our data.
Participants were informed that webcam proctoring was applied to create a standardized testing situation and to ensure that the test was completed independently and without impermissible aids. The rules were stated in general terms rather than as an exhaustive list of specific permissible and impermissible behaviors. A standardized cheating protocol was developed that included clear indicators of cheating specific to our remote testing environment (e.g., discussing solutions with another person, another person providing an answer unsolicited, taking a photo of the screen with a cell phone, using a cell phone to solve a task, consulting a cheat sheet, writing notes, or self-disclosing one’s own cheating). Additionally, we recorded other relevant events that might impact the participant or the testing situation but could not be clearly classified as cheating (e.g., leaving the testing area, background noise, wearing headphones). All predefined options for cheating and other relevant events were available for selection in separate drop-down menus in an excel sheet. If none of the predefined options were applicable, or if additional information was necessary for evaluation, the observer could enter text in a designated field. For each event, the exact start time within the video was also documented.
The video recordings were reviewed by a trained student assistant and one co-author, each of whom was responsible for approximately half of the recordings. To prepare and obtain an initial overview, four video recordings were evaluated by both raters on a trial basis, followed by a detailed briefing and joint discussion. To determine the consistency between the raters, the video recordings of ten randomly selected participants were viewed by both raters. Cheating and non-cheating participants were identified with 100% consistency. The identification of other relevant events with major impact also showed perfect consistency across raters.
For each participant a separate digital cheating protocol was created. The video recordings, which typically consisted of two separate parts per participant (i.e., the first and second part of the intelligence test), were initially played at normal speed and then accelerated to eight times the speed within the first minute. At eightfold speed, audio output was not available, so only the visual stream could be used to identify relevant events. If irregular or notable events were detected (e.g., a participant appeared to be speaking), the video was paused, and the relevant segment was reviewed at normal speed. Sometimes (see below) the video stream, audio output, or segments of the video were missing. In such cases, the assessment was limited to the available material, and when only audio was present, the playback was slowed accordingly.

2.3.3. Main Analyses

Measurement Invariance
We conducted a series of confirmatory factor analyses (CFAs) to evaluate configural, metric, and scalar levels of measurement invariance, using a stepwise approach and testing each intelligence ability separately.
Configural measurement invariance was tested by modeling each intelligence ability as a one-factor model, with each ability represented by its respective indicators (i.e., test tasks). Figure 2 illustrates, as an example, the modeling of the reasoning factor by its nine manifest indicators. As each indicator also reflected either numerical, verbal, or figural content, we allowed the residual variances of items sharing the same content to correlate, as it is plausible according to the BIS-model that these items share unique variance. A test scale of an ability was considered as configural invariant if the model demonstrated adequate fit in all proctoring conditions. Specifically, we interpreted a CFI > 0.95, an RMSEA < 0.06, an SRMR < 0.08, and a non-significant χ2-test as indicative of good model fit (Hu and Bentler 1999). Furthermore, CFI > 0.90, RMSEA < 0.08, and SRMR < 0.10 were considered as a minimally acceptable fit. Note that in some instances, it was necessary to modify the model by allowing additional residual correlations or by removing residual correlations between items sharing the same content, due to issues with non-convergence or unacceptable fit indices across all conditions. In such cases, we explicitly reported the modifications and modeled them across all conditions.
To test metric measurement invariance, we first examined a baseline model assuming configural invariance by specifying the previously established CFA models as a multigroup model across the proctoring conditions. Metric measurement invariance was then evaluated by constraining the factor loadings to be equal across conditions and comparing the fit indices of this model to the baseline model. Following Chen (2007), changes in CFI (ΔCFI) of 0.01 or less were interpreted as indicating negligible deterioration in model fit.
Next, scalar measurement invariance was tested by constraining the intercepts of the indicators to be equal across conditions. Additionally, the mean of the latent factor for respective ability was fixed to zero in the proctored condition and estimated freely in the other conditions. The fit of the scalar model was compared to the metric model using the same fit indices and criteria. If scalar invariance was established, it was possible to compare the means of the latent factors across conditions. In this case, significant deviations from zero for the means in the UP random or UP chosen conditions would indicate that the latent means differ significantly from those in the proctored condition. If full measurement invariance could not be established, we assessed partial measurement invariance. Following the guidelines of Dimitrov (2010), we allowed up to 20% of the parameters to vary across groups, as this threshold is considered acceptable for establishing partial invariance.
Test Performance Mean Differences
To test mean differences in test performance across proctoring conditions and ability tests, we conducted a repeated measures MANOVA with proctoring condition (proctored, UP random, UP chosen) as the between-subjects factor and ability test (reasoning, short-term memory, processing speed, divergent thinking) as the within-subjects factor. The main effects of the condition and ability test as well as their interaction were tested. Tukey’s HSD tests were conducted as post hoc comparisons to further examine differences in factor scores between conditions for each ability test.
Test Duration
The test duration was analyzed using one-way ANOVAs with proctoring condition as the predictor. As the test was administered in two parts with a short break in between, which could differ in length between participants, we conducted the analysis separately for each part.
Frequency of Cheating
Proctoring records were descriptively analyzed to determine how many participants in the proctored condition could be identified as cheaters. In addition, the specific types of cheating employed by these participants were documented. Other relevant events that could have interfered with the testing situation were also descriptively analyzed.
Robustness Analyses
All main analyses were repeated after excluding participants from the proctored condition who were identified as cheaters based on video review as a robustness check.

3. Results

3.1. Preliminary Results

Frequency distributions for gender, current employment or student status, highest education, parental education, and first language by condition are reported in Table 4. χ2-tests indicated no significant differences between proctoring conditions on gender, parents’ education, and primary language (p > 0.05). However, χ2-tests indicated significant differences in employment status (χ2 = 18.477, df = 10, p = 0.047) and highest education (χ2 = 19.472, df = 10, p = 0.035). Participants in the UP random condition were less frequently full-time employed (28%) and more frequently university students (40%) compared to the Proctored condition (full-time: 47%, university: 31%) and UP chosen condition (full-time: 42%, university: 29%). Among proctored participants, the highest educational degree was more often a university degree (46%) and less often a high school diploma (Abitur; 12%) or intermediate school degree (11%), relative to the UP random (university degree: 21%, high school diploma: 20%, intermediate degree: 15%) and UP chosen conditions (university degree: 32%, high school diploma: 19%, intermediate degree: 19%). Moreover, UP random condition participants more often had a lower secondary school degree as their highest education (17%) than participants in the other two conditions (Proctored condition: 8%, UP chosen condition: 9%). However, ANOVA results (F = 0.152, df = 2, p = 0.859) showed that the proctoring conditions did not differ with respect to age (Proctored condition: M = 29.34, SD = 5.84; UP random condition: M = 29.56, SD = 6.51; UP chosen condition: M = 29.91, SD = 6.89).

3.2. Main Results

3.2.1. Measurement Invariance

Model fit indices of all CFAs and measurement invariance tests are reported in Table 5. For reasoning, a one-factor model indicated a good fit for the proctored and UP random conditions and an acceptable model fit for the UP chosen condition after some modifications (for more detail, see Table 5), indicating configural measurement invariance. Metric and scalar measurement invariances could also be established.
For short-term memory, a one-factor model indicated a good fit for the proctored and UP chosen conditions and an acceptable model fit for the UP random condition after some modifications (for more detail, see Table 5), indicating configural measurement invariance. Metric measurement invariance was narrowly achieved (ΔCFI = 0.010), whereas scalar measurement invariance was not achieved (ΔCFI = 0.043). Partial scalar measurement invariance could only be achieved when 6 of 27 intercepts were freely estimated, which exceeds the 20% threshold deemed acceptable by Dimitrov (2010).
In detail, the following modifications had to be performed to achieve partial scalar invariance. For M_f3 (i.e., the third memory indicator referring to figural content), the proctored group intercept was freely estimated, whereas the UP random and UP chosen group intercepts were constrained to be equal. The proctored group had an intercept of 0.197 (p = 0.065), the UP random group had an intercept of −0.184 (p = 0.156), and the UP chosen group had an intercept of −0.014 (p = 0.893). For M_n1 (i.e., the first memory indicator referring to numerical content), the UP chosen group intercept was freely estimated, whereas the proctored and UP random group intercepts were constrained to be equal. The proctored group had an intercept of −0.333 (p = 0.001), the UP random group had an intercept of 0.035 (p = 0.773), and the UP chosen group had an intercept of 0.281 (p = 0.012). For M_v1 (i.e., the first memory indicator referring to verbal content), the UP chosen group intercept was freely estimated, whereas the proctored and UP random group intercepts were constrained to be equal. The proctored group had an intercept of −0.110 (p = 0.213), the UP random group had an intercept of −0.153 (p = 0.218), and the UP chosen group had an intercept of 0.254 (p = 0.037). For M_v2 (i.e., the second memory indicator referring to verbal content), the proctored group intercept was freely estimated, whereas the UP random and UP chosen group intercepts were constrained to be equal. The proctored group had an intercept of −0.286 (p = 0.003), the UP random group had an intercept of 0.094 (p = 0.456), and the UP chosen group had an intercept of 0.175 (p = 0.127). To sum up, across the four relevant test indicators, two (M_n1 and M_v2) showed lower intercepts for the proctored group compared to both UP groups, one (M_f3) showed lower intercepts for both UP groups compared to the proctored group, and one (M_v1) showed a mixed pattern, with the UP random group scoring slightly lower and the UP chosen group scoring higher than the proctored group.
For processing speed, we observed configural, metric, and scalar measurement invariance without needing any modifications.
Finally, in divergent thinking, configural measurement invariance could be demonstrated after some modifications (for more detail, see Table 5) through very good model fits across proctoring conditions. Metric and scalar measurement invariance could also be established (ΔCFI < 0.01).
Latent factor scores for reasoning, processing speed, and divergent thinking were saved for further analyses as manifest values from multigroup scalar measurement invariance models across conditions. The factor scores for short-term memory were obtained from a multigroup partial scalar invariant model and were also saved for further analyses.

3.2.2. Test Performance Mean Differences

Mean differences in factor scores across conditions are depicted in Figure 3, and the exact means and standard deviations are provided in Table 6. There was no significant main effect of condition on factor scores (F(2, 223) = 1.29, p = 0.276). The within-subject factor ability test showed a significant effect (F(3, 221) = 8.96, p < 0.001). Because the factor mean for proctored participants was fixed at zero and the factor scores for the UP random and UP chosen conditions were freely estimated, this result indicates that the groups deviated from the reference category (proctored) in different ways across the various ability tests. Furthermore, the interaction between condition and ability test was significant (F(6, 444) = 3.06, p = 0.006). This indicates that the effect of proctoring condition on test performance differed by ability test. As can be seen in Figure 3, proctored participants outperformed those in the UP random and UP chosen conditions in reasoning and processing speed, whereas they performed worse in memory. For divergent thinking, participants in the UP chosen condition showed lower test scores compared to those in the proctored or UP random conditions. However, although the interaction effect was significant, Tukey post hoc tests revealed no significant differences between factor scores across conditions for any ability except for divergent thinking, where proctored participants scored significantly higher than those in the UP chosen condition (p = 0.046).
As an additional exploratory analysis, we repeated the same ANOVA with condition as the between-subjects factor and ability test as the within-subjects factor, this time including employment status and highest education, as well as all possible interactions, as additional factors. This was performed to test whether the results remained robust when accounting for a priori group differences reported in our preliminary analyses. In this ANOVA, employment status was not a significant predictor (F(5, 203) = 2.24, p = 0.052), while highest education was significant (F(4, 203) = 4.69, p = 0.001). However, the overall results pattern for condition and test type remained unchanged. There was no significant main effect of condition (F(2, 223) = 1.15, p = 0.318), a significant main effect of the within-subject factor ability test (F(3, 221) = 10.22, p < 0.001), and a significant interaction between condition and ability test (F(6, 444) = 3.27, p = 0.004).

3.2.3. Test Duration

With respect to test duration, ANOVAs revealed no significant differences in mean duration across conditions in both the first (F = 1.461, df = 2, p = 0.234) and the second (F = 2.386, df = 2, p = 0.095) parts of the test (proctored condition: M1/M2 = 88.52/37.66, SD1/SD2 = 15.46/5.46; UP random: M1/M2 = 84.43/35.20, SD1/SD2 = 9.90/6.85; UP chosen: M1/M2 = 87.75/39.72, SD1/SD2 = 19.49/19.12).

3.2.4. Frequency of Cheating

Of the 74 participants in the proctored condition, video recordings were available for all participants. Among these 74 participants, we identified three clear cases of cheating (4%). The detected cheating behaviors included taking cell phone photos of the screen, using a cell phone to solve tasks, writing down notes as a cheat sheet, and reading solutions aloud from the cell phone. In all three cases, the cheating behavior was clearly identifiable. For example, one participant took photos of the screen with a mobile phone in a mostly concealed manner, visible only when the phone was briefly raised above desk level.
Other relevant events, such as leaving the testing area, the presence of non-participating individuals, or background noises (e.g., music), occurred several times. However, such events were only classified as a major source of interference if they stood out from other participants in terms of quality or quantity (e.g., leaving the room for about ten minutes, or a combination of street noise, background music, gardening, and short conversations). Major technical issues, such as missing both video and audio output, were also categorized as relevant events with significant impact. In total, we identified 11 participants (15%) with noticeable sources of interference compared to the rest of the sample.

3.2.5. Robustness Analyses

Replicating the complete main analyses after excluding the data of the three cheating participants from the proctored condition revealed no changes in the established level of measurement invariance or in latent factor score differences between the proctoring conditions, with one negligible exception. For the short-term memory test, only partial metric invariance was achieved, as one factor loading had to be freely estimated across proctoring conditions, whereas full metric invariance was established in the full sample.

4. Discussion

In the present study, we investigated the effects of proctoring with webcam video recordings of intelligence tests, focusing on four intelligence abilities: reasoning, short-term memory, processing speed, and divergent thinking. First, we examined measurement invariance across three proctoring conditions (proctored, UP random, and UP chosen) for each ability test separately. CFAs revealed scalar measurement invariance for reasoning, processing speed, and divergent thinking, indicating that mean-levels of these abilities are comparable across conditions. Short-term memory, however, did not show scalar measurement invariance due to substantial intercept differences across conditions. This lack of scalar invariance implies that the memory tasks did not measure the same latent construct equivalently across groups, limiting comparability.
Second, we investigated whether mean test performance differed across proctoring conditions and whether this effect varied by ability test. There was no significant main effect of proctoring condition, indicating that proctoring did not systematically increase or decrease test performance across the different abilities. However, we found a significant interaction between condition and ability test, partly in accordance with our hypotheses. Descriptively, proctored participants performed somewhat worse in short-term memory but slightly better in reasoning and processing speed compared to participants in both unproctored conditions. For divergent thinking, only participants in the UP chosen condition showed lower performance relative to the proctored condition. Nevertheless, observed differences within each ability test were of small effect size and only the difference in divergent thinking between the proctored and UP chosen conditions reached statistical significance in post hoc comparisons. Thus, despite some indication that proctoring may influence test performance, the overall effects were limited and varied by assessed ability.
Third, we investigated differences in test duration across conditions. Contrary to assumptions about cheating affecting test-taking time, no significant differences in test duration were found.
Fourth, reviewing video recordings revealed a low prevalence of observable cheating behaviors (4%) under proctored conditions, aligning with prior findings suggesting that explicit cheating remains relatively uncommon even in remote assessments. The types of observed cheating behaviors, such as writing notes or photographing the screen, are strategies typically associated with memory-based tasks, where scalar measurement invariance was not established.

4.1. Strengths and Limitations

Before discussing the implications of our findings for proctoring research and remote testing practice, we acknowledge strengths and limitations of the present study. A notable strength is the diverse assessment of abilities, including reasoning, short-term memory, processing speed, and divergent thinking. However, comprehension-knowledge, another important cognitive domain, was not assessed. Investigating comprehension-knowledge in future research is particularly relevant, as the cheatability of typical comprehension-knowledge tests might be especially high in unproctored remote settings due to internet access. An additional strength of the present study is the explicit assessment of measurement invariance across proctoring conditions. Indeed, our analyses demonstrated the importance of measurement invariance, as scalar invariance could not be established for memory. Therefore, in our analyses of mean differences across conditions and ability tests, we used factor scores from a partial scalar invariance model for memory, in which some intercepts were allowed to vary between conditions. It should be noted as a limitation that we slightly exceeded the commonly recommended threshold for partial measurement invariance as proposed by Dimitrov, which suggests freeing no more than 20% of the intercepts. In our model, 6 out of 27 intercepts (i.e., 22%) had to be freely estimated to achieve partial measurement invariance. Similarly to Wright et al. (2014), the pattern of group differences in item intercepts that were not invariant was not consistent across the indicators. While two items showed lower intercepts for the proctored group and one item showed lower intercepts for both unproctored groups, another item displayed a mixed pattern, with the UP random group showing lower and the UP chosen group showing a higher intercept than the proctored group. Moreover, it should be noted that cheating in unproctored settings is likely limited to a relatively small subsample of participants. While our measurement invariance approach captures group-level differences in test functioning across conditions, it may be less sensitive to effects that occur only in a small number of individuals. Future research could complement group-level measurement invariance approach with person-centered analyses, which may be more sensitive in detecting such subgroup effects.
Another strength of our study is the diverse sample, comprising participants aged 20 to 44 with variability in professional status and educational attainment. The experimental assignment between the proctored condition and the UP random condition ensured random allocation between these groups. However, we nevertheless observed differences between conditions in employment status and highest education. These differences may provide alternative explanations for our findings and we therefore conducted additional ANOVAs including employment status and highest education as factors. These analyses did not alter the observed effects: condition remained a non-significant between-person factor, while a significant interaction between condition and ability test persisted.
The relatively small sample size limited statistical power for detecting small mean-level differences, which possibly also explains why post hoc tests did not reveal significant differences between the conditions in any individual test despite the presence of a significant interaction of condition and test type (with the exception of divergent thinking). Additionally, the small sample size prevented the estimation of more complex CFA models incorporating all abilities simultaneously. With a larger sample, it would have been possible to investigate structural invariance among abilities, such as potential differences in the relationships between reasoning and memory across proctoring conditions. Our sample size may also have contributed to convergence issues in some models, necessitating modifications such as removing or allowing correlations between residual variances in our analyses.
Finally, the intelligence testing in this study was conducted under low stake conditions, meaning participants were neither substantially rewarded nor penalized based on their performance. Thus, participants had little extrinsic incentive to engage in cheating. Furthermore, in the present study, participants received a fixed payment regardless of performance, which may have encouraged some to complete tasks quickly and “well enough” rather than with maximal accuracy, potentially reducing the motivation to cheat. To mitigate this risk, we informed participants that their data would be checked for signs of serious engagement before payment approval. Of note, one previous study also reported cheating in low-stakes conditions, suggesting that cheating is not limited to high-stakes testing (Karim et al. 2014, see Table 2). Although this limits generalizability to contexts involving high stake conditions (e.g., employment or university applications), our study realistically reflects common research scenarios typically characterized by low stake conditions, thus ensuring ecological validity within this context.

4.2. Implications and Future Research

One important finding of this study is that we observed scalar measurement invariance for all assessed ability tests except for short-term memory, suggesting that memory tasks functioned differently across proctoring conditions. Our review of recorded videos indicated that cheating behaviors primarily involved participants photographing their screens or noting down answers, which are particularly beneficial for memory tasks. If such cheating occurred in the two unproctored conditions, this could explain why scalar measurement invariance was not established for the memory test. The specific nature of the observed cheating behavior is noteworthy, as previous studies have attributed task-specific differences across proctoring conditions mainly to the ease of searchability (e.g., Karim et al. 2014), which facilitates cheating in unproctored settings. In contrast, our findings suggest that it was primarily the opportunity for note-taking, rather than searchability alone, that impacted the measurement properties and comparability of memory tasks across different proctoring conditions. Consequently, our findings suggest that memory tasks should not be administered without proctoring, as cheating may be particularly easy and influential in these tasks. It is also important to be careful with other tasks where taking notes could help, such as mental arithmetic problems.
In addition to searchability and note-taking, future research should also address the potential role of AI solvability in intelligence tests. Our current study found no evidence of AI-related cheating. Such cheating might have been suspected if the measurement invariance tests had revealed differences in verbal reasoning or divergent thinking tasks across proctoring conditions. However, ongoing advancements in AI and its integration into everyday applications, such as browser-based AI support, could facilitate cheating in the future. For example, advances in AI tools could mean that not only knowledge-based or vocabulary items but in some cases also more complex reasoning or problem-solving tasks might become solvable through AI assistance.
The findings regarding mean performance differences across proctoring conditions and abilities have several important implications. The absence of a significant main effect of proctoring suggests that remote testing without proctoring might generally provide comparable results to proctored testing, at least in low-stakes research settings. This is consistent with previous research on intelligence testing (Ihme et al. 2009; Lievens and Burke 2011; Norrøne and Nordmo 2025; Schakel 2012; Templer and Lange 2008), which has primarily focused on reasoning tests. Thus, both our study and most prior research on proctoring effects in intelligence testing contradict the findings of Steger et al.’s (2018) meta-analysis, which was based primarily on studies conducted in educational contexts and which reported that unproctored participants outperformed those in proctored conditions. This suggests that such effects may not generalize to intelligence testing.
However, the significant interaction between condition and ability test highlights that proctoring effects may vary by ability, and importantly, differences may not always disadvantage proctored participants. In our analysis, we observed descriptive advantages for proctored participants in reasoning and processing speed, as well as a significant advantage in divergent thinking compared to the UP chosen group. Similar performance benefits under proctored conditions have previously been reported (e.g., Steger et al. 2018; Alessio et al. 2018) and could be explained by motivational factors or increased concentration in supervised settings. As the observed differences across proctoring conditions varied across ability tests, it is possible that the motivational impact of webcam monitoring differs across ability domains, with some tasks that rely more heavily on effort or concentration benefiting more than others. The idea is also in line with some previous findings of superior performance in proctored conditions for tests of low cheatability (Coyne et al. 2005; Williamson et al. 2017; see Table 2).
This suggests that researchers should exercise caution when interpreting the results of unproctored tests, even for tasks that are difficult to cheat at, as participants may not exert maximum effort or concentration in unsupervised conditions.
Furthermore, we found no differences in test duration between the proctoring conditions, which argues against the occurrence of extensive cheating in the unproctored condition, as such behavior would likely have resulted in longer test durations. One possible explanation for these findings is that cheating in most of the tasks used in our study (i.e., reasoning, processing speed, and divergent thinking) is unlikely, as the solutions are not easily searchable or amenable to note-taking. Alternatively, the finding of no differences in test duration between the proctoring conditions could be explained by a very low amount of cheating overall. As expected, within the proctored condition, we identified only a very small proportion of cheaters (4%), which is in line with previous research on intelligence tests that also found little or no evidence of cheating (Lievens and Burke 2011; Norrøne and Nordmo 2025; Wright et al. 2014). Note that in the present study, webcam-based proctoring involved only video recording of the participant without screen capture. While this approach can effectively detect overt behaviors such as note-taking or photographing the screen, it is less capable of identifying covert behaviors, including online searches or the use of AI-assisted tools. Other forms of proctoring such as live remote monitoring, screen recording, or in-person supervision may offer broader coverage of potential cheating behaviors, but they also come with higher costs, greater intrusiveness, and possible privacy concerns.
To conclude, cheating appears to happen rarely in low-stakes settings when using complex intelligence tests, and we are therefore generally optimistic about the validity of intelligence testing in online remote environments. However, when memory tasks are included, we strongly recommend the use of proctoring, as cheating cannot be ruled out due to the ease of note-taking. The same caution applies to tasks that are highly searchable (Steger et al. 2018), such as knowledge questions in comprehension-knowledge tests. Moreover, advantages for proctored participants, potentially due to increased motivation or higher concentration, should also be considered when choosing the testing environment. Therefore, if proctoring is feasible, we generally recommend its implementation. For complex tasks, such as reasoning or divergent thinking, or tasks where cheating is inherently difficult, such as processing speed, we remain optimistic that cheating is rare and negligible even without proctoring. Finally, when comparing proctored and unproctored conditions, we strongly recommend conducting measurement invariance tests before comparing test means across groups. This is crucial, as tests may differ not only in their mean scores across proctoring conditions, but also in whether the same indicators measure the underlying ability construct in a comparable way. Overall, researchers and practitioners should carefully consider test-specific vulnerabilities when choosing between proctored and unproctored online cognitive assessments.

Author Contributions

Conceptualization, V.S., M.B. and F.P.; methodology, V.S.; software, V.S.; validation, V.S., N.P., M.B., J.U., J.P. and F.P.; formal analysis, V.S. and J.U.; investigation, V.S., N.P., M.B.; resources, F.P.; data curation, V.S. and J.P.; writing—original draft preparation, V.S., N.P. and M.B.; writing—review and editing, V.S., N.P., M.B., J.U., J.P. and F.P.; visualization, V.S. and J.U.; supervision, F.P.; project administration, V.S., M.B. and F.P.; funding acquisition, V.S., M.B. and F.P.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hogrefe Publishing House.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Ethics Committee of Trier University (protocol code: 46/2025; date of approval: 24 April 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Analysis code and dataset are available online on: https://osf.io/yznct/?view_only=66c66d3627104459926aaf4afa265c74 (Accessed on 26 August 2025).

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT-5 and ChatGPT-4o for linguistic refinements and for generating repetitive code segments. These tools were not used for generative purposes related to the substantive content of the manuscript. The authors have reviewed and edited all outputs and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Alessio, Helaine M, Nancy Malay, Karsten Maurer, A. John Bailer, and Beth Rubin. 2018. Interaction of proctoring and student major on online test performance. The International Review of Research in Open and Distributed Learning 19: 165–85. [Google Scholar] [CrossRef]
  2. Baso, Yusring Sanusi. 2022. Proctoring and non-proctoring systems. International Journal of Advanced Computer Science and Applications 13: 75–82. [Google Scholar] [CrossRef]
  3. Borsboom, Denny. 2006. When Does Measurement Invariance Matter? Medical Care 44: S176–81. [Google Scholar] [CrossRef] [PubMed]
  4. Chan, Jason C. K., and Dahwi Ahn. 2023. Unproctored online exams provide meaningful assessment of student learning. Proceedings of the National Academy of Sciences of the United States of America 120: e2302020120. [Google Scholar] [CrossRef]
  5. Chen, Binglin, Sushmita Azad, Max Fowler, Matthew West, and Craig Zilles. 2020. Learning to cheat: Quantifying changes in score advantage of unproctored assessments over time. In Proceedings of the Seventh ACM Conference on Learning @ Scale. New York: Association for Computing Machinery, pp. 197–206. [Google Scholar] [CrossRef]
  6. Chen, Fang Fang. 2007. Sensitivity of goodness of fit indexes to lack of measurement invariance. Structural Equation Modeling a Multidisciplinary Journal 14: 464–504. [Google Scholar] [CrossRef]
  7. Cheung, Gordon W., and Roger B. Rensvold. 2002. Evaluating goodness-of-fit indexes for testing measurement invariance. Structural Equation Modeling 9: 233–55. [Google Scholar] [CrossRef]
  8. Coyne, Iain, Tim Warszta, Sarah Beadle, and Noreen Sheehan. 2005. The impact of mode of administration on the equivalence of a test battery: A quasi-experimental design. International Journal of Selection and Assessment 13: 220–24. [Google Scholar] [CrossRef]
  9. Daffin, Lee William, Jr., and Ashley Anne Jones. 2018. Comparing student performance on proctored and non-proctored exams in online psychology courses. Online Learning Journal 22: 131–45. [Google Scholar] [CrossRef]
  10. Dendir, Seife, and R. Stockton Maxwell. 2020. Cheating in online courses: Evidence from online proctoring. Computers in Human Behavior Reports 2: 100033. [Google Scholar] [CrossRef]
  11. Dimitrov, Dimiter M. 2010. Testing for factorial invariance in the context of construct validation. Measurement and Evaluation in Counseling and Development 43: 121–49. [Google Scholar] [CrossRef]
  12. Domínguez, César, Javier López-Cuadrado, Anaje Armendariz, Arturo Jaime, Jónathan Heras, and Tomás A. Pérez. 2019. Exploring the differences between low-stakes proctored and unproctored language testing using an internet-based application. Computer Assisted Language Learning 32: 483–509. [Google Scholar] [CrossRef]
  13. Gohel, David. 2023. officer: Manipulation of Microsoft Word and PowerPoint Documents 0.6.6. Available online: https://cran.r-project.org/package=officer (accessed on 26 August 2025).
  14. Gohel, David, and Panagiotis Skintzos. 2017. flextable: Functions for Tabular Reporting 0.9.6. Available online: https://doi.org/10.32614/cran.package.flextable (accessed on 26 August 2025).
  15. Feinman, Lena. 2018. Alternative to Proctoring in Introductory Statistics Community College Courses. Publication No. 4622. Walden Dissertations and Doctoral Studies. Ph.D. thesis, Walden University, Minneapolis, MN, USA. Available online: https://scholarworks.waldenu.edu/dissertations/4622 (accessed on 26 August 2025).
  16. Fletcher, Richard, and John Hattie. 2011. Intelligence and Intelligence Testing. Abingdon-on-Thames: Routledge. [Google Scholar]
  17. Fox, John, Sanford Weisberg, and Brad Price. 2023. car: Companion to Applied Regression 3.1.2. Available online: https://CRAN.R-project.org/package=car (accessed on 26 August 2025).
  18. Henry, Lionel, and Hadley Wickham. 2023. purrr: Functional Programming Tools 1.0.2. Available online: https://CRAN.R-project.org/package=purrr (accessed on 26 August 2025).
  19. Hu, Li-tze, and Peter M. Bentler. 1999. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling a Multidisciplinary Journal 6: 1–55. [Google Scholar] [CrossRef]
  20. Ihme, Jan Marten, Franziska Lemke, Kerstin Lieder, Franka Martin, Jonas C. Müller, and Sabine Schmidt. 2009. Comparison of ability tests administered online and in the laboratory. Behavior Research Methods 41: 1183–89. [Google Scholar] [CrossRef]
  21. Jäger, Adolf O. 1984. Intelligenzstrukturforschung: Konkurrierende Modelle, neue Entwicklungen, Perspektiven [Research on intelligence structure: Competing models, new developments, perspectives]. Psychologische Rundschau 35: 21–35. [Google Scholar]
  22. Jäger, Adolf O., Heinz Holling, Franzis Preckel, Ralf Schulze, Miriam Vock, Heinz M. Süß, and Andre Beauducel. 2006. Berliner Intelligenzstruktur-Test für Jugendliche: Begabungs- und Hochbegabungsdiagnostik (BIS-HB). Göttingen: Hogrefe. [Google Scholar]
  23. Karim, Michael N., Samuel E. Kaminsky, and Tara S. Behrend. 2014. Cheating, reactions, and performance in remotely proctored testing: An exploratory experimental study. Journal of Business and Psychology 29: 555–72. [Google Scholar] [CrossRef]
  24. Kingston, Neal, and Amy Clark, eds. 2014. Test Fraud: Statistical Detection and Methodology. Abingdon-on-Thames: Routledge. [Google Scholar]
  25. Langenfeld, Thomas. 2020. Internet-Based Proctored Assessment: Security and Fairness Issues. Educational Measurement: Issues and Practice 39: 24–27. [Google Scholar] [CrossRef]
  26. Lievens, Filip, and Eugene Burke. 2011. Dealing with the threats inherent in unproctored internet testing of cognitive ability: Results from a large-scale operational test program. Journal of Occupational and Organizational Psychology 84: 817–24. [Google Scholar] [CrossRef]
  27. Little, Todd D. 1997. Mean and Covariance Structures (MACS) Analyses of Cross-Cultural Data: Practical and Theoretical issues. Multivariate Behavioral Research 32: 53–76. [Google Scholar] [CrossRef]
  28. Mackintosh, Nicholas J. 1998. IQ and Human Intelligence. Oxford: Oxford University Press. [Google Scholar]
  29. Mohammed, Hussein M., and Qutaiba I. Ali. 2022. Cheating prevention in e-proctoring systems using secure exam browsers: A case study. Jurnal Ilmiah Teknik Elektro Komputer dan Informatika (JITEKI) 8: 634–48. [Google Scholar] [CrossRef]
  30. Nettelbeck, Ted, and Carlene Wilson. 2005. Intelligence and IQ: What teachers should know. Educational Psychology 25: 609–30. [Google Scholar] [CrossRef]
  31. Norrøne, Tore Nøttestad, and Morten Nordmo. 2025. Comparing proctored and unproctored cognitive ability testing in high-stakes personnel selection. International Journal of Selection and Assessment 33: e70001. [Google Scholar] [CrossRef]
  32. Oeding, Jill, Theresa Gunn, and Jamie Seitz. 2024a. The impact of remote online proctoring versus no proctoring: A study of graduate courses. Online Journal of Distance Learning Administration 27: 4. Available online: https://ojdla.com/articles/the-impact-of-remote-online-proctoring-versus-no-proctoring-a-study-of-graduate-courses (accessed on 26 August 2025).
  33. Oeding, Jill, Theresa Gunn, and Jamie Seitz. 2024b. The mixed-bag impact of online proctoring software in undergraduate courses. Open Praxis 16: 82–93. [Google Scholar] [CrossRef]
  34. Ones, Deniz S., Chockalingam Viswesvaran, and Stephan Dilchert. 2017. Cognitive ability in personnel selection decisions. In The Blackwell Handbook of Personnel Selection. Edited by Arne Evers, Neil Anderson and Olga Voskuijl. Hoboken: Blackwell Publishing, pp. 143–73. [Google Scholar] [CrossRef]
  35. Öncül, Bilal. 2021. Dealing with cheating in online exams: A systematic review of proctored and non-proctored exams. International Technology and Education Journal 5: 45–54. Available online: https://itejournal.com/en/dealing-with-cheating-in-online-exams-a-systematic-review-of-proctored-and-non-proctored-exams.htm (accessed on 26 August 2025).
  36. R Core Team. 2024. _R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing. Available online: https://www.R-project.org/ (accessed on 26 August 2025).
  37. Reisenwitz, Timothy H. 2020. Examining the necessity of proctoring online exams. Journal of Higher Education Theory and Practice 20: 118–24. [Google Scholar] [CrossRef]
  38. Revelle, William. 2023. psych: Procedures for Psychological, Psychometric, and Personality Research 2.4.3. Available online: https://cran.r-project.org/web/packages/psych/index.html (accessed on 26 August 2025).
  39. Rodríguez-Villalobos, Martha, Jessica Fernandez-Garza, and Yolanda Heredia-Escorza. 2023. Monitoring methods and student performance in distance education exams. The International Journal of Information and Learning Technology. 40, pp. 164–76. [CrossRef]
  40. Rosseel, Yves. 2012. lavaan: An R Package for Structural Equation Modeling. Journal of Statistical Software 48: 1–36. [Google Scholar] [CrossRef]
  41. Schakel, Lolle. 2012. Online Computer-Based Testing in Human Resource Management: Contributions from Item Response Theory. University of Groningen/UMCG research database (Pure). Ph.D. thesis, University of Groningen, Groningen, The Netherlands. Available online: https://research.rug.nl/en/publications/online-computer-based-testing-in-human-resource-management-contri (accessed on 26 August 2025).
  42. Schmitt, Neal, and Goran Kuljanin. 2008. Measurement invariance: Review of practice and implications. Human Resource Management Review 18: 210–22. [Google Scholar] [CrossRef]
  43. Stanley, David. 2021. apaTables: Create American Psychological Association (APA) StyleTables 2.0.8, R package version. Available online: https://cran.r-project.org/web/packages/apaTables/index.html (accessed on 26 August 2025).
  44. Steger, Diana, Ulrich Schroeders, and Timo Gnambs. 2018. A meta-analysis of test scores in proctored and unproctored ability assessments. European Journal of Psychological Assessment 36: 174–84. [Google Scholar] [CrossRef]
  45. Tatum, Holly, and Beth M. Schwartz. 2017. Honor codes: Evidence based strategies for improving academic integrity. Theory into Practice 56: 129–35. [Google Scholar] [CrossRef]
  46. Taylor, John L., William R. Lindsay, and Paul Willner. 2008. CBT for people with intellectual disabilities: Emerging evidence, cognitive ability and IQ effects. Behavioural and Cognitive Psychotherapy 36: 723–33. [Google Scholar] [CrossRef]
  47. Templer, Klaus J., and Stefan R. Lange. 2008. Internet testing: Equivalence between proctored lab and unproctored field conditions. Computers in Human Behavior 24: 1216–28. [Google Scholar] [CrossRef]
  48. Wickham, Hadley. 2024. ggplot2: Create Elegant Data Visualisations Using the Grammarof Graphics 3.5.1. Available online: https://CRAN.R-project.org/package=ggplot2 (accessed on 26 August 2025).
  49. Wickham, Hadley, Jim Hester, Romain Francois, Jennifer Bryan, Sam Bearrows, Jouni Jylänki, and Mikkel Jørgensen. 2023. Package “readr”: Read Rectangular Text Data 2.1.5. Available online: https://cran.r-project.org/web/packages/readr/readr.pdf (accessed on 26 August 2025).
  50. Wickham, Hadley, Romain François, Lionel Henry, Kirill Müller, and Davis Vaughan. 2014. dplyr: A Grammar of Data Manipulation. Available online: https://doi.org/10.32614/cran.package.dplyr (accessed on 26 August 2025).
  51. Williamson, Kenneth C., Vickie M. Williamson, and Scott R. Hinze. 2017. Administering spatial and cognitive instruments in-class and on-line: Are these equivalent? Journal of Science Education and Technology 26: 12–23. [Google Scholar] [CrossRef]
  52. Wright, Natalie A., Adam W. Meade, and Sara L. Gutierrez. 2014. Using invariance to examine cheating in unproctored ability tests. International Journal of Selection and Assessment 22: 12–22. [Google Scholar] [CrossRef]
  53. Zhang, Niu, James Larose, and Megan Franklin. 2024. Effect of unproctored versus proctored examinations on student performance and long-term retention of knowledge. The Journal of Chiropractic Education 38: 114–19. [Google Scholar] [CrossRef]
  54. Zhao, Li, Junjie Peng, Shiqi Ke, and Kang Lee. 2024. Effectiveness of unproctored vs. teacher-proctored exams in reducing students’ cheating: A double-blind randomized controlled field experimental study. Educational Psychology Review 36: 126. [Google Scholar] [CrossRef]
Figure 1. BIS+ indicators based on the berlin intelligence structure model. Note. r = reasoning, s = processing speed, m = short-term memory, dt = divergent thinking, f = figural ability, v = verbal ability, n = numerical ability.
Figure 1. BIS+ indicators based on the berlin intelligence structure model. Note. r = reasoning, s = processing speed, m = short-term memory, dt = divergent thinking, f = figural ability, v = verbal ability, n = numerical ability.
Jintelligence 13 00110 g001
Figure 2. Confirmatory factor analysis of resoning based on nine content specific indicators. Note. f = figural ability, v = verbal ability, n = numerical ability. Each indicator represents reasoning ability and refers to either figural, verbal, or numerical content. Residual correlations between indicators with the same content domain were admitted.
Figure 2. Confirmatory factor analysis of resoning based on nine content specific indicators. Note. f = figural ability, v = verbal ability, n = numerical ability. Each indicator represents reasoning ability and refers to either figural, verbal, or numerical content. Residual correlations between indicators with the same content domain were admitted.
Jintelligence 13 00110 g002
Figure 3. Mean differences in intelligence factor scores by proctoring condition. Note. P. Speed = processing speed. D. Thinking = divergent thinking. All factor mean scores except for short-term memory stemmed from scalar measurement invariance multigroup confirmatory factor analyses with proctoring condition as the multigroup factor. Short-term memory factor scores stemmed from the partial scalar measurement invariant multigroup confirmatory factor analysis.
Figure 3. Mean differences in intelligence factor scores by proctoring condition. Note. P. Speed = processing speed. D. Thinking = divergent thinking. All factor mean scores except for short-term memory stemmed from scalar measurement invariance multigroup confirmatory factor analyses with proctoring condition as the multigroup factor. Short-term memory factor scores stemmed from the partial scalar measurement invariant multigroup confirmatory factor analysis.
Jintelligence 13 00110 g003
Table 1. Overview of recent studies testing the effects of proctoring.
Table 1. Overview of recent studies testing the effects of proctoring.
Authors (Year)Assessment Type or ParadigmProctoring SolutionDesignNMean AgeAge RangePopulationCountryStakesAverage PerformanceCheatingTest DurationMeasurement Invariance
Alessio et al. (2018)academic testingsoftwarewithin97 university studentsUShighunproctored > proctorednot testedunproctored > proctorednot tested
Baso (2022)academic testingsoftwarewithin101 19–21university studentsIndonesiahighunproctored > proctorednot testednot testednot tested
Chan and Ahn (2023)academic testingon-site proctoringwithin2010 university studentsUShighno differenceyes (ineffective at boosting performance)not testednot tested
Chen et al. (2020)academic testingon-site proctoringwithin510 university studentsUShighunproctored > proctoredyes (increasing over the semester)not testednot tested
Daffin and Jones (2018)academic testingvideo monitoringwithin1694 mostly 30–50university students (with employment)UShighunproctored > proctorednot testedunproctored > proctorednot tested
Dendir and Maxwell (2020)academic testingvideo recording between, not random64819.99–21.40 university studentsUShighunproctored > proctoredyes (prior to proctoring)not testednot tested
Domínguez et al. (2019)academic testingon-site proctoringbetween, not random158423.52 18–39general populationSpainlowunproctored > proctorednot testedno difference not tested
Feinman (2018)academic testingon-site proctoringwithin85026 university studentsUShighSet 1: no difference
Set 2: proctored > unproctored
not testednot testednot tested
Norrøne and Nordmo (2025)cognitive ability testingon-site proctoringwithin4872119–27Norwegian Armed ForcesNorwayhighno differenceyes (minimal)not testedStrict
Oeding et al. (2024a)academic testingsoftwarebetween, not random426 university studentsUShighunproctored > proctorednot testednot testednot tested
Oeding et al. (2024b)academic testingsoftwarebetween, not random252 university studentsUShighmixednot testednot testednot tested
Reisenwitz (2020)academic testingsoftware; on-site proctoringbetween, not random72 university studentsUShighunproctored > proctorednot testednot testednot tested
Rodríguez-Villalobos et al. (2023)academic testingsoftware; on-site proctoringbetween, not random296 25–63university students (postgraduates)Mexicohighunproctored > proctorednot testednot testednot tested
Zhang et al. (2024)academic testingon-site proctoringbetween, not random234 university students USSample 1: high, Sample 2: lowSample 1: unproctored > proctored
Sample 2: proctored > unproctored
not testedunproctored > proctorednot tested
Zhao et al. (2024)fake question paradigmon-site proctoringbetween, randomized177/158/191/166 19–22university studentsChinahighnot testedproctored > unproctorednot testednot tested
Table 2. Overview of studies investigating the effects of proctoring in intelligence test settings.
Table 2. Overview of studies investigating the effects of proctoring in intelligence test settings.
Authors (Year)Cognitive TestsProctoring SolutionDesignNMean AgeAge RangePopulationCountryStakesAverage PerformanceCheatingTest DurationMeasurement Invariance
Coyne et al. (2005)numerical, verbal, and figural reasoningon-site proctoringwithin86/
64
31.51/
22.77
professionals/university studentsUKlowproctored > unproctorednot testednot testednot tested
Ihme et al. (2009)figural matriceson-site proctoringbetween, not random48121.7/
24.7/
27.0
18–46/
18–50/
18–56
mixedGermanylowno differencenot testednot testednot tested
Karim et al. (2014)Graduate Record Examination
test/
Raven’s matrices
video recordingbetween, randomized29530 general populationUSlowsearchable:
unproctored > proctored
not searchable:
no difference
unproctored > proctorednot testednot tested
Lievens and Burke (2011)numerical and verbal reasoningon-site proctoringwithinca. 3500 Mostly 21–29applicantsUKhighno differencelow frequency of aberrant scores (0.3–2.2%)not testednot tested
Norrøne and Nordmo (2025)figure matrices, word similarities, number reasoning/mental rotationon-site proctoringwithin4872119–27Norwegian Armed ForcesNorwayhighno differenceyes (minimal) strict
Schakel (2012)figure series, matrices, number serieson-site proctoringwithin42527 applicantsNetherlandshighno differencemedium frequency of aberrant scores (5–12%) not tested
Templer and Lange (2008)creativityon-site proctoringwithin and between (randomized)16321.919–25potential applicantsSingaporelowno difference
Williamson et al. (2017)spatial ability, reasoningon-site proctoringbetween, randomized45719.6418–32university studentsUSAlowspatial: mixed
reasoning:
proctored > unproctored
not testednot testednot tested
Wright et al. (2014)deductive reasoningsoftwarebetween, not randomized24,750/
9615
applicantsUSAhighunproctored > proctored/
proctored > unproctored
no evidence of cheatingproctored > unproctored/
unproctored > proctored
differential item functioning in 8 of 30 indicators
Table 3. Description of the assessed intelligence ability scales.
Table 3. Description of the assessed intelligence ability scales.
AbilityDescriptionExample Task
ReasoningTo understand new information, identify connections, and draw conclusions through thoughtful analysis.Solving figural analogies.
Short-Term MemoryTo quickly memorize information and then recall or recognize it correctly.Memorize two-digit numbers and freely recall them.
Processing SpeedTo stay focused and work quickly and accurately on simple tasks that require little deliberation.Deciding whether a word (e.g., oak) is part of a category (e.g., tree) or not.
Divergent ThinkingTo generate many original ideas by approaching problems from different perspectives.Naming abilities a salesman should not have.
Table 4. Frequency Distribution by Condition.
Table 4. Frequency Distribution by Condition.
VariableCategoryProctoredUn-ProctoredNo Camχ2dfp
GenderFemale19/26%23/31%33/43%5.70740.222
Male53/72%51/68%43/56%
Non-binary2/3%1/1%1/1%
Employment StatusFull-time
employed
35/47%21/28%32/42%18.477100.047
Part-time
employed
9/12%6/8%8/10%
Self-employed1/1%7/9%4/5%
Unemployed6/8%4/5%3/4%
University student23/31%30/40%22/29%
Others0/0%7/9%7/9%
Highest EducationHigh school
diploma (Abitur)
9/12%15/20%15/19%19.472100.035
Intermediate school degree8/11%11/15%15/19%
Lower secondary school6/8%13/17%7/9%
No school degree15/20%13/17%13/17%
University degree34/46%16/21%25/32%
Others2/3%5/7%0/0%
Parents EducationDoctorate (PhD)4/5%5/7%3/4%11.679120.307
High school
diploma (Abitur)
17/23%12/16%12/16%
Intermediate school degree16/22%18/24%22/29%
Lower secondary school7/9%7/9%16/21%
Others2/3%2/3%4/5%
University degree28/38%30/40%19/25%
Primal LanguageGerman63/85%64/85%64/83%0.17520.916
Others11/15%11/15%13/17%
Table 5. Fit indices from confirmatory factor analyses and measurement invariance testing.
Table 5. Fit indices from confirmatory factor analyses and measurement invariance testing.
Modelnχ2SFdfpCFIRMSEASRMRΔχ2ΔdfΔpΔCFIΔ RMSEAΔ SRMRAICBIC
Reasoning Ability
Configural MI
Proctored 1 7422.8430.969190.2440.9790.0500.044 16711752
UP random 1 7519.3020.945190.438100.041 17991880
UP chosen 1 7731.6750.956190.0340.9250.0960.061 18101892
Baseline22673.860.957570.0660.970.0600.049 52815640
Metric MI22691.5590.989730.0700.9620.0600.08718.013160.323−0.00800.03852695573
Scalar MI226102.2750.992890.1590.9730.0460.09210.843160.8190.011−0.0140.00552475497
Memory Ability
Configural MI
Proctored 27417.750.977150.2760.9880.0440.043 15641654
UP random 2 7522.5881.033150.0930.9610.080.055 18581949
UP chosen 2776.9841.031150.958100.0340010.039−0.08−0.02117941885
Baseline22647.2341.014450.3810.9990.0150.04440.472300.096−0.0010.0150.01052175617
Metric MI22667.8311.013610.2560.9890.0370.08120.608160.194−0.0100.0220.03752065551
Scalar MI226106.4121.005770.0150.9460.0710.09939.232160.001−0.0430.0340.01852125503
P. Scalar MI 322681.1341.016700.1710.9800.0450.08313.23890.152−0.0090.0080.00252015516
Processing Speed Ability
Configural MI
Proctored 7423.8750.927180.1590.9810.0690.040 14441527
UP random 7525.8070.987180.1040.9680.0930.047 15961679
UP chosen 7720.6430.901180.2980.9930.0420.0360010.025−0.051−0.01116371722
Baseline22670.5540.938540.0650.9810.0710.04149.744360.063−0.0120.0290.00546785047
Metric MI22696.1010.924700.0210.9710.0750.08725.797160.057−0.010.0040.04646684983
Scalar MI226109.2720.941860.0460.9740.0650.09013.815160.6120.003−0.010.00346514910
Divergent Thinking Ability
Configural MI
Proctored 4 740.3970.9810.529100.014 912956
UP random 4 751.1630.79310.281100.022 9931037
UP chosen 4 772.3580.8910.1250.9780.120.029001−0.0220.120.00710181062
Baseline2263.8420.88830.279100.0221.4820.4770.022−0.12−0.00729243119
Metric MI22610.3431.011110.500100.0496.66480.573000.02729153082
Scalar MI22615.2410.987190.707100.0584.80780.778000.00929033043
Note. MI = measurement invariance. UP = unproctored. 1 = Models were modified by removing one residual correlation between two indicators referring to figural content. 2 = Models were modified by allowing three additional residual correlations suggested by modification indices. 3 = Partial scalar MI was achieved after allowing six of 27 intercepts to vary between proctoring conditions. 4 = Models were modified by allowing three additional residual correlations suggested by modification indices.
Table 6. Latent correlations among intelligence abilities and their mean scores across proctoring conditions.
Table 6. Latent correlations among intelligence abilities and their mean scores across proctoring conditions.
VariableProctoredUP RandomUP Chosen123
MSDMSDMSD
1. Reasoning0.000.51−0.210.66−0.190.55
2. Memory−0.010.610.060.710.160.590.41 **
[0.29, 0.51]
3. Processing Speed−0.010.74−0.320.89−0.240.860.71 **0.40 **
[0.64, 0.77][0.28, 0.50]
4. Divergent Thinking0.000.28−0.010.29−0.120.330.35 **0.20 **0.44 **
[0.23, 0.46][0.08, 0.33][0.33, 0.54]
Note. ** p < 0.01. Values in square brackets indicate the 95% confidence interval for each correlation. Latent factor scores for reasoning, processing speed, and divergent thinking were obtained from multigroup scalar measurement invariant models across conditions. The factor scores for memory were obtained from a multigroup partial scalar invariance model. In these models, the respective factor score for the proctored condition was fixed at 0, whereas the factor scores for the UP random and UP chosen conditions were estimated freely.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Scherrer, V.; Petry, N.; Breit, M.; Urban, J.; Preuß, J.; Preckel, F. Effects of Proctoring on Online Intelligence Measurement: A Literature Overview and an Empirical Study. J. Intell. 2025, 13, 110. https://doi.org/10.3390/jintelligence13090110

AMA Style

Scherrer V, Petry N, Breit M, Urban J, Preuß J, Preckel F. Effects of Proctoring on Online Intelligence Measurement: A Literature Overview and an Empirical Study. Journal of Intelligence. 2025; 13(9):110. https://doi.org/10.3390/jintelligence13090110

Chicago/Turabian Style

Scherrer, Vsevolod, Nicolai Petry, Moritz Breit, Julian Urban, Julian Preuß, and Franzis Preckel. 2025. "Effects of Proctoring on Online Intelligence Measurement: A Literature Overview and an Empirical Study" Journal of Intelligence 13, no. 9: 110. https://doi.org/10.3390/jintelligence13090110

APA Style

Scherrer, V., Petry, N., Breit, M., Urban, J., Preuß, J., & Preckel, F. (2025). Effects of Proctoring on Online Intelligence Measurement: A Literature Overview and an Empirical Study. Journal of Intelligence, 13(9), 110. https://doi.org/10.3390/jintelligence13090110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop