1. Introduction
Since Boeing started developing the first digital avatar with a human image in the early 1960s, this technology has primarily moved from the laboratory or research setting into the commercial sector. It is believed to have become increasingly important in the digital age (
Cui and Liu 2023). From a technical point of view, virtual humans are virtual individuals who have become a reality in recent years due to the advancement of information technology, specifically in sectors like motion capture, computer graphics, machine learning, high-precision rendering, and others. The early digital human image generation models were inefficient and expensive and mostly created with modelling software. On the other hand, more accurate features or human pictures based on 3D Morphable Models have been made possible using deep learning algorithms and data-driven 3D reconstructions (
Sela et al. 2017;
Tewari et al. 2017;
Tran and Liu 2018). Developments in deep learning have led to the widespread adoption of end-to-end neural network approaches for speech synthesis, which has significantly improved the quality of speech. Prior systems primarily relied on joint rules and language-specific knowledge; however, advances in speech synthesis have also been critical to improving the accuracy of these representations (
Ren et al. 2019;
Zhao et al. 2019;
Zhu et al. 2019;
Yang et al. 2020;
Wu et al. 2021). As the technology continues to evolve, the concept is said to have expanded from just being a research tool, and it has seen increased integration, with various use cases and cross-disciplinary capabilities.
The virtual human market was estimated to be around USD 11.30 billion in 2013 and will reach USD 440.3 billion soon, close to a 44.7% annual growth rate (
Gandham 2022;
Dharmadhikari 2024). This growth has been seen as a result of the pandemic, during which technologies like the metaverse impacted our outlook and changed the online world to make the use of virtual spaces acceptable, which led to an increasing shift towards digitalisation across all sectors. This growth has been attributed to the fast-paced development observed with technology, and it also divides the market into two categories: 2D virtual humans and 3D virtual humans. Two-dimensional virtual humans are flat representations in 2D animations and games, while 3D ones are more realistic and immersive, and often used within virtual reality applications. Virtual humans have seen increased adoption in the entertainment industry, which has aimed at creating digital characters within the service industry for customer service and virtual assistants. In the education industry, they are used for simulations and training and also find a prominent presence in other sectors like marketing, healthcare, and virtual events (
Allied Market Research 2023).
There are growing concerns regarding their adoption, as technology is often considered a disruptor that could be used positively or negatively (
ITU 2019). While modern technology has allowed for increased connectivity, it has led to the increased use of social media and mobile devices, which leads to eye strain, short attention spans, and depression (
Twenge and Campbell 2018). Virtual humans have been used recently as a way of connecting with dead loved ones, where people seek help from AI-generated avatars to process their grief after the death of a beloved family member. This is achieved using deepfake technology, a diffusion model that can develop a realistic avatar that can move and speak. China is said to be a critical market when it comes to this technology, and the sector developing AI avatars has matured at a fast pace and has been profiting from developing this technology. Though the technology offers a short-term solace to individuals who have lost loved ones, there are several ethical considerations associated with its development and application. These include the practicality of the technology and the potential for data exploitation (
Yang 2024). There are many technical challenges with this even now, like the challenges of replicating or cloning body movements or certain facial expressions, and most clients often do not have enough high-quality data for the result to be satisfactory. Companies like SuperBrain rely on permission from family members. Still, a company in Ningbo recently used AI tools to create videos of celebrities who have passed away using public data and without approval (
Yang 2024). This technology has also allowed for increased scamming, with people being able to replicate voices and communicate with a persons’ close friend, leading to trust being broken as they often are fooled by virtual humans (
Visser 2024). The increased possibility of the usage of technology to defraud people has also increased distrust of the said technology.
Research on virtual humans and their perception has been a significant area of interest, and this research has focused on exploring voices that lead to more confidence among people when communicating with digital humans. A study mentioned that while voice quality did not impact learning, it did indeed influence trust and learners’ perception of virtual humans. This finding is critical to understanding how trust is built and how actively this factor influences the situation (
Chiou et al. 2020). It is mentioned that even at the time of the study, there were no established measurement tools for trust, especially for virtual humans in virtual reality. A few studies have identified different tools, like the socially conditioned place preference paradigm (SCPP), as ones that could help with measurements. Still, a study has highlighted the importance and role trust plays when it comes to accepting and adopting new technology (
Cronje 2023). Researchers found that some or most participants thought virtual humans were more trustworthy in trustworthy conditions than those considered untrustworthy. The study involved asking participants to follow or ask for advice from virtual humans, which was thought to be a behavioural measure of trust (
Schroeder et al. 2021). Importantly, manipulation during the research impacts trust behaviour; those in trustworthy conditions would be more likely to ask for advice and follow it, which indicates a paradigm overly sensitive to the assessment of interpersonal trust towards virtual humans (
Lin et al. 2023).
This research focuses on acceptance and trust but also explores a region not covered well in earlier research. This is the impact of distinct environments on acceptance, which includes social media, everyday life interactions, and business settings. In some fields, such as entertainment and social media, digital humans are considered critical and often supported due to the innovation and unique experiences they offer (
Hwang and Hwang 2022;
Deng and Jiang 2023;
Ham et al. 2024). The transparency regarding their virtual nature tends to increase engagement and also develops a positive perception of them. Still, this acceptance is said to be different when it comes to the apprehension and scepticism observed in the business environment and in professional and personal interactions (
Glass 2024;
Moon et al. 2024). The potential issues of digital humans in these contexts raise significant ethical and security concerns, overshadowing their benefits. This engenders a degree of distrust, and by examining their initial positive response and subsequent changes in attitude on the revelation of their trusting nature, this study aims to understand the underlying factors influencing people’s trust and acceptance of virtual humans. This understanding can help developers and marketers optimise these agents’ deployment and assist policymakers in effectively regulating their use.
This research is critical, as with the growing importance and prominence of digital humans across different sectors, understanding the dynamics of their interactions becomes essential and also helps to provide insight into the societal readiness for widespread AI integration, identifying the potential barriers that exist when it comes to their acceptance and proposing strategies that would increase engagement while addressing ethical concerns. This study’s outcomes help develop a guide for developers, regulators, and businesses to make informed decisions that align with people’s expectations and ethical standards.
3. Materials and Methods
This study aimed to investigate how the perception of virtual humans has evolved and to examine people’s behavioural changes before and after the disclosure of their virtual nature. This research employed a mixed-methods approach, combining quantitative and qualitative analyses. The quantitative component utilised standardised survey instruments to measure changes in perceptions and attitudes, allowing for objective, replicable data collection that could be generalised to a broader population. This study examined the impact of virtual human disclosures across three contexts: social media, business/professional environments, and general public settings. This study included participants across three age groups—18–25 years, 26–40 years, and 40–60 years—with a distribution of 31%, 35%, and 24%, respectively. Participants younger than 18 and older than 60 were excluded to maintain consistency in the sample. Despite the inclusion of different age groups, the analysis did not reveal any significant differences in the perception and acceptance of virtual humans based on age. This finding suggests that the age-related factors influencing AI acceptance may not be as pronounced within the studied age range, or that the differences were too subtle to detect with the given sample size. It is possible that the relatively tech-savvy nature of the younger participants, as well as the professional experience of the older age group, contributed to a more uniform response across the board. However, the lack of significant age-related differences contrasts with some past research, which has suggested that younger individuals are generally more accepting of AI and digital innovations. This discrepancy may be due to the specific contexts in which the virtual humans were used in this study, which might have resonated similarly across age groups.
3.1. Experiment Design
The social media experiment focused on assessing the impact of disclosing the virtual nature of a communication counterpart on participants’ perceptions. Participants interacted with video content featuring a virtual human (VH) presented as a social media influencer. Engagement metrics, such as liking, sharing, and commenting, were recorded without the participants’ knowledge that the influencer was virtual. Following this interaction, participants completed a survey evaluating their initial impressions, including the VH’s overall appeal, trustworthiness, and engagement. In the disclosure phase, participants were informed that the influencer was a virtual human, and detailed information about the underlying technology and its typical uses in marketing was provided. Subsequently, a follow-up survey assessed changes in perceived authenticity, trust, and willingness to engage with the VH in the future. The study employed a within-subject before-and-after design, where participants’ perceptions were measured immediately before and after the disclosure of the virtual human’s nature. This design inherently controls for individual differences, as each participant serves as their own control. The lack of a separate control group is justified, given the study’s focus on immediate perceptual changes due to this disclosure, with no significant time gap between the two measurement points. This approach minimises the influence of external variables and provides a direct assessment of the impact of the disclosure.
The purpose of the business/professional environment was to evaluate how people interacted with virtual agents and how exposing the usage of virtual humans affected people’s perceptions in a work context. The first task required of the participants was to imitate business contacts with an agent they saw as human. These interactions resembled common business situations such as client meetings and customer support. The survey was used to evaluate the agents’ professionalism, efficiency, and the respondent’s level of satisfaction following their contact. Similar to the previous experiment, the participants were also told that the agent was a VH, and they were given comprehensive details on the limits and powers of virtual agents in a work environment. Following this, the same poll was utilised to gauge how their view had changed after learning these details. The third experiment was about general public perception, which focused on societal trust and ethical concerns regarding the use of VHs in everyday scenarios. The participants encountered virtual humans in a simulated scenario from their personal life, communicating with the customer care service of a bank, all the while unaware that they were interacting with a VH. A survey with different questions was used to collect their perception in these conditions, which looked mostly at the performance and trustworthiness of virtual humans. After completing this step, the individuals were informed about the VHs they engaged with and given additional information about the technologies used, including their challenges and benefits. They were provided with a lot of information, which covered topics such as the potential use of technologies like deep-fake video creation and how they are utilised in scams. The discussion also touched on the challenges and ethical considerations of the technology, highlighting the current difficulties in its adoption while emphasising its overall benefits. Once the information was provided and people were ready, they were asked to take another survey based on how their views had changed with the information provided.
To create the interaction or simulation of an interaction, there was a need to develop virtual humans that would be realistic, and for this we used Microsoft’s VASA-1 (
Figure 1), which works by using AI, machine learning, computer graphics, and natural language processing, which creates all the required technology and has the expertise required for creating a realistic human with behavioural adaptations and efficient text-to-speech synthesis (
Xu et al. 2024). Multiple agents or VHs were developed and programmed to operate in different environments based on the need to conduct experiments set in different scenarios. Microsoft provided the needed VHs according to our testing specifications.
3.2. Data Collection
This study employed a rigorous sampling strategy and a detailed data collection process to ensure the reliability and validity of its findings. Between December 2023 and February 2024, 371 participants from the United Kingdom were recruited across three experiments. The study involved three distinct experiments, each conducted in different contexts: social media (Experiment 1), a business/professional environment (Experiment 2), and in a general public setting (Experiment 3). The number of participants varied slightly across these experiments due to the voluntary nature of their participation and minor attrition. Specifically, Experiment 1 included 130 participants, Experiment 2 included 125 participants, and Experiment 3 included 118 participants. These variations in participant numbers resulted in corresponding differences in the number of observations for each experiment. This analysis accounts for these differences by focusing on within-subject comparisons using a before-and-after design. Each experiment randomly selected 10 participants for follow-up interviews to gain deeper insights into the changes in their perceptions. The sample was gender-balanced, with 190 female and 181 male participants ranging in age from 18 to 65 years. The analysis indicated no significant demographic effects on the outcomes. Participants were recruited using a snowball sampling method, beginning within the Cambridge college community.
Once the interactions were completed, in each phase, a survey was distributed, which was focused on pre-disclosure reactions. The survey was very structured and used a six-point Likert scale to gauge the participant’s positive or negative reaction or perception of the technology specific to the component of the study they were in. The survey covers multiple areas of engagement, such as appeal, efficiency, professionalism, credibility, satisfaction, and others. Once this was completed, educational materials were provided, which helped the participants understand the technology and the potential benefits and risks. A second survey was distributed to the same group to collect responses in this post-declaration phase. In all the components, two open-ended questions reflected the change in the people’s opinion before and after the disclosure.
3.3. Data Analysis
The questions from the surveys used for the three experiments were analysed and grouped into four factors or themes. The four factors identified were trust, performance, usage likelihood, and overall acceptance. The scores for these factors were calculated as the average of the scores given for the questions in each group.
Table 1 provides a list of the questions (see
Appendix B as well) grouped under each factor, and these too were divided into experiments and based on whether they were answered before or after the disclosure (TBEX1, TAE1 and so on), so that separate analyses could be carried out on the data from each experiment and then their overall score.
The collected data were subjected to paired t-tests to assess the differences between the pre- and post-disclosure scores for each survey question, enabling an analysis of shifts in perception. The t-test statistics determined whether the means of the two groups (pre- and post-disclosure) differed significantly. A p-value of less than 0.05 was considered to indicate a statistically significant relationship, suggesting that the observed differences were unlikely to be due to random chance. A large t-statistic value would show a significant difference or stronger effect of the disclosure. This helped with the comparison at the individual scale and, as a combination, crossed the three experiments in terms of the answers given before and after the disclosure and helped us to see in which situation or context the shift was greater and in which context there were fewer variations. In addition to calculating the paired t-test values and p-values, Cohen’s d was computed to determine the effect size of the difference between pre- and post-disclosure measures. This approach allows for a clearer interpretation of the magnitude of the change in participants’ trust, performance, usage likelihood, and overall acceptance of VHs. This research also used a regression analysis, descriptive statistics, and Q-Q plots to better understand the difference in participants’ values or perceptions before and after the disclosure.
Responses to open-ended questions and data from the 30 interviews were analysed using a thematic analysis to explore participants’ opinions and attitudes toward virtual humans after their disclosure. This qualitative analysis aimed to identify and understand the underlying reasons for the shifts in perception, with recurring themes highlighted and presented in the results.
3.4. Ethical Consideration
All participants were informed that the study might involve an initial deception, but they were assured that they would be fully debriefed after the interaction. During the debriefing, the purpose of the deception was explained, and participants were provided with complete information about the study’s objectives and procedures. Their rights and any emotional or ethical concerns they might have also been addressed. The participants were not coerced into participating in the experiments. They volunteered for the study and were informed about the requirements and how it would be beneficial. They were also given the option to withdraw from the research if they felt uncomfortable. All participants have provided written informed consent.
4. Results and Discussion
It is important to note that the interpretation of the trust scores varies depending on the experimental context. In Experiment 1 (social media), a higher score indicates a higher level of trust. However, in Experiment 3 (general public perception), a higher score reflects lower trust. This distinction arises due to the use of different scales for the trust measurement in each context. In the following sections, the direction of the trust scores will be clearly indicated to avoid confusion.
A comprehensive list of the abbreviations used in this chapter can be found in
Appendix A.
The study involved 130 participants in Experiment 1, 125 in Experiment 2, and 118 in Experiment 3. The analysis began with an examination of descriptive statistics, providing a foundation for comparing the shifts in the key variables across the three experimental contexts. We then categorised the data and their corresponding answers based on four criteria. After that, we calculated the overall response shift across the studies based on three factors.
Table 2 presents descriptive statistics for trust, performance, usage likelihood, and overall acceptance across all three experiments. The number of observations (n) in each experiment is as follows: Experiment 1 (n = 130), Experiment 2 (n = 125), and Experiment 3 (n = 118).
Table 2 highlights significant variations in trust before and after the disclosure of the nature of the virtual human (VH) across three experimental contexts. In Experiment 1 (social media context), the mean trust score slightly decreased from 2.1487 (positive) to 1.9795, indicating a mild reduction in trust. In this experiment, higher scores indicate more trust. This shift suggests that although the overall attitude remained within a positive range, the revelation of the VH’s nature led to a slight drop in confidence.
In Experiment 2 (business/professional environment), an interesting trend is observed, where trust increased post-disclosure, from 1.1290 to 1.3044. This change suggests that transparency in professional settings might enhance trust, as participants might value the honesty about the VH’s nature. Despite this increase, trust remains in the extremely positive range, indicating a generally favourable perception.
Conversely, Experiment 3 (general public perception) reveals a substantial decline in trust from 1.0684 to 2.7179 after the disclosure, showcasing a shift from an extremely positive to a moderately positive perception. In this context, a higher score corresponds to lower trust. Therefore, this increase reflects a significant decline in trust after the disclosure. This sharp decline indicates that the general public is more sceptical and less accepting of VHs once their non-human nature is disclosed, leading to a significant drop in trust.
Regarding performance, Experiment 1 shows a slight increase from 1.3538 to 1.4712 after the disclosure, remaining within the extremely positive range. This increase suggests that participants in the social media context may appreciate the technical capabilities of VHs, even after learning about their artificial nature. In Experiment 2, the performance perception increased marginally, from 1.8360 to 1.9839, indicating that participants still perceived the VH’s performance positively, though not as strongly so as before. Experiment 3 exhibited a slight decline in performance from 1.0983 to 1.6937, yet it remained within the extremely positive range. This suggests that, despite the drop, their performance is still viewed favourably.
Usage likelihood presents a more pronounced shift. In Experiment 1, the likelihood decreased significantly, from 2.5846 to 1.9077, showing a move from a positive perception closer to a slightly positive outlook post-disclosure. Similarly, in Experiment 2, the usage likelihood shifted from 1.4194 (extremely positive) to 1.8871, indicating a reduction but still remaining within a generally positive range. However, in Experiment 3, there was a dramatic increase from 1.0812 to 4.6496, shifting from an extremely positive to a nearly negative perception. This indicates a significant drop in people’s willingness to engage with VHs after their nature is revealed, reflecting a substantial reduction in their acceptance in the general public context.
Finally, overall acceptance reflects these trends across all three experiments. In Experiment 1, VHs’ acceptance slightly decreased from 2.1526 to 1.9833, remaining within the positive range. Experiment 2 shows a minor decline from 1.7961 to 2.0737, indicating that their overall acceptance remains positive but is slightly less enthusiastic post-disclosure. Experiment 3 shows a more significant drop from 1.1154 to 2.3533, moving towards a more neutral outlook, suggesting a considerable reduction in the overall acceptance of VHs in the general public context.
These results indicate that while some participants maintain a positive perception of VHs after their disclosure, particularly in professional contexts, there is a notable trend towards reduced trust, performance perception, usage likelihood, and overall acceptance, especially within general public settings. The findings underscore the importance of context in shaping attitudes towards VHs, with transparency playing a critical role in influencing trust and acceptance.
The paired
t-test of the Experiment 3 (
Table 3) results indicates statistically significant shifts in all variables post-disclosure, with the largest mean difference observed in the usage likelihood (−1.10,
p < 0.001), suggesting a substantial decrease in participants’ willingness to engage with virtual humans after learning of their artificial nature. While the study examined perceptions of virtual humans across three distinct contexts—social media, business environments, and general public interactions—the results from Experiment 3 (general public perception) are presented here as representative of broader societal trends. The general public context is particularly critical in understanding societal acceptance and the ethical concerns surrounding virtual humans, as it provides the most comprehensive and generalisable insights. Furthermore, the patterns of trust, performance, and usage likelihood observed in Experiment 3 are consistent with the trends seen in the other two experiments, allowing for a focused presentation of the most significant and impactful findings without redundancy.
The paired
t-test for the trust results before and after disclosure shows a mean difference of −0.5978813, indicating that trust decreased on average by this amount after the disclosure. The t-statistic of −11.09816 reflects the size of this difference relative to the variability in the data. However, rather than relying solely on the t-statistic, Cohen’s d was calculated to provide a standardised measure of the effect size. For Experiment 3, Cohen’s d was found to be 1.65, indicating a large effect and a substantial decline in trust following the disclosure. This significant shift suggests that the disclosure of the virtual human’s nature had a strong impact on trust. The
p-value of 0.00 is below the commonly accepted threshold of 0.05, indicating that the change in trust before and after the disclosure is statistically significant and unlikely to be due to chance. This statistical significance, combined with the large effect size (Cohen’s d = 1.65), further emphasises the substantial impact of the disclosure on public trust. Across all four factors, the pre-disclosure values are consistently lower than the post-disclosure values, reflecting a shift in attitudes after the virtual human’s nature was revealed. Performance before and after disclosure had a mean difference of −0.2931356, with a t-statistic of −13.8722. The corresponding Cohen’s d value is 0.59, indicating a medium effect size. Usage likelihood had the largest mean difference of −1.100932, with a t-statistic of −11.7349. Cohen’s d was calculated to be 3.57, indicating a very large effect size and a significant reduction in participants’ willingness to use virtual humans post-disclosure. The data reveal that all pre- and post-disclosure comparisons are statistically significant, with pre-disclosure values consistently lower across the board. This shift indicates a move towards more negative perceptions of virtual humans following their disclosure, especially in terms of trust and usage likelihood. The mean difference in overall acceptance was −0.5594, with a t-statistic of −10.1593. The corresponding Cohen’s d value is 1.23, indicating a large effect size and a notable reduction in overall acceptance following the disclosure. The Q-Q plots for trust and usage acceptance (
Figure 2 and
Figure 3) reveal deviations from normality, particularly in post-disclosure scores. However, the Cohen’s d values, which were found to be large in several instances (e.g., trust, d = 1.65; usage likelihood, d = 3.57), provide a robust indication of the effect sizes that are significant, even in the presence of non-normal distributions. This non-normal distribution may have influenced the paired
t-test results, with more pronounced deviations observed in the post-disclosure data. The curve of the points suggests differing distributions between pre- and post-disclosure, with pre-disclosure values being less spread out and more tightly clustered.
This study then examined each of the experiments or components individually to determine which had the greatest impact. We started by examining how trust changes with the experiments or contextual situations (
Table 4).
The trust levels exhibited the most significant decline in the general public experiment (mean trust before: 1.07, mean trust after: 2.71,
p < 0.001), indicating that participants were most sceptical about virtual humans in everyday scenarios. The first experiment is truly a unique setting, as it is the only one where trust has shifted more towards positive than all the others, which the t-statistic confirms. The results indicated that prior to disclosure, the perception of trust was higher. However, after disclosure, the value decreased or shifted closer to a positive direction within the context of social media. This suggests an increased trust in virtual humans (VHs) on social media, consistent with the existing literature’s findings. On the other hand, the t-statistics were negative. After disclosure, the perception scores increased, shifting towards the negative side. In Experiment 2, it was −5.5016, while in Experiment 3, it was −17.8707, showing an increasing difference or negative perception regarding trust in these two scenarios. For the first trust experiment result, we used a Q-Q plot (
Figure 4) and observed that the lower values of the variables are quite similar, while there is a significant difference between the higher values.
The use of descriptive statistics allows for a better understanding of the variables, particularly in terms of their ranges. In Experiment 1, there was a high level of variation among the variables. Initially, the level of trust was low even before the nature of the virtual human was known. Following the revelations, the trust values decreased even further, with the highest value being 4.667 across all factors. In the second experiment (
Table 5) the difference between the maximum values was only 0.5. The maximum value increased from 1.75 to 2.25. However, in the third experiment, there was a significant drop in trust, as the maximum value changed from 2 to 4. Based on the literature, it is understood that social media is an area in which change is anticipated, and it is a platform that people trust more. However, in other areas such as business and everyday life, there has been a more highly negative perception once trust is established. While the mean value alone does not give a complete picture, descriptive statistics can elucidate the range of this perception.
After analysing the performance of VHs in the three experiments, it was observed that there was a decrease in performance perception across the different experiments after the reveal. Notably, the third experiment (
Table 6) exhibited a higher negative t-statistic value of 14.3827, indicating a significant difference in the means relative to the overall variability in the data. Similarly, Experiment 2 also shows a significantly high difference. In contrast, in Experiment 1, the values are lower compared to the other two, which shows that revealing the VH led to an increased shift towards the negative among the respondents. The largest shift was observed in the general public perception experiment, followed by the business environment experiment. For Experiment 1 (social media environment), the negative perception could highlight increased expectations of the quality or performance of VHs.
In
Figure 5, the Q-Q plot indicates that the lower values are similar, but the higher values after the revelations were found to be larger in their deviation than the before values. This is shown by the points below the line, which demonstrate significant divergence as the value increases.
The descriptive statistics in
Table 7 reveal interesting findings, particularly in relation to the range, minimum, and maximum values. In the first experiment, there was a shift in perception, with maximum values of 2.25 before and 3 after, indicating a shift of 0.75 towards the negative. On the other hand, in the second experiment, the shift was even smaller, with a change of just 0.33 between the minimum and maximum acceptance levels. The third experiment showed a significant perception shift, with values changing from 1.67 before to 3 after the revelations, indicating a reduced perception of performance in general public perception tests.
Focusing on usage likelihood, we can see that on social media, the usage likelihood has an increased positive effect among the participants, with the mean dropping from 2.5846 to 1.9076, while the t-statistic has a value of 3.4775, which means the shift is towards a more positive perception. In the case of business and general public perceptions, the situation was different. In the business context of Experiment 2, the shift was comparatively lower but shifted towards the negative. The post-disclosure mean value of perception was 4.64, indicating a higher level of negative perception or a lower chance of future usage. The t-statistic was also −19.2796, revealing significant differences in its values before and after. This suggests an increase in the likelihood of social media VH usage, but a decrease in the likelihood of its usage in the other two scenarios.
The paired
t-test results (
Table 8) for usage likelihood reveal significant differences before and after the VH reveal across all examples. For social media (EX1), usage likelihood decreased from a mean of 2.58 to 1.91, with a statistically significant t-statistic of 3.48 and
p-value of 0.0007, indicating a decrease in usage intent. In business contexts (EX2), usage likelihood increased from 1.42 to 1.89, supported by a t-statistic of −6.11 and a highly significant
p-value (<0.001). For general usage (EX3), a substantial increase in likelihood was observed, from 1.08 to 4.65, with a t-statistic of −19.28 and a
p-value also less than 0.001.
The Q-Q plot of Experiment 3 (
Figure 6) indicates that the higher values deviated significantly from the line. These points could be considered outliers or significant changes. This suggests that after Experiment 3, people’s views have significantly altered, and their likelihood of using VHs is lower.
When we look at the descriptive statistics of usage likelihood (
Table 9), we notice that the minimum and maximum values for social media have not changed much, but the average value has. This indicates that there are still many people who have a negative perception of the use of virtual humans in social media, but there is a significantly larger population that might prefer it. On the other hand, in the business and general perception experiments, we observed a significant shift in the minimum and maximum values. This suggests that the introduction of virtual humans has a negative impact and alters people’s perception of the likelihood of their use.
The final factor is overall acceptance (
Table 10). Similar to the previous factors, their overall acceptance in social media improves after the re-evaluation. However, this is in contrast to the other two factors, where there is a significant decrease in their acceptance after the revelation. This suggests that people are more cautious about using technology in these situations.
In previous tests, we have observed that the higher values for overall acceptance in the general perception experiment, or Experiment 3, vary significantly from the values found before the revelation. This shows a shift towards a more negative attitude or perception (
Figure 7).
The descriptive statistics (
Table 11) align with observations made that are related to other factors. The business and general perception experiments have shown a significant increase in the maximum values from before to after the reveal. In the context of social media, while the maximum value has increased, the values in the middle have either dropped or shifted in the positive direction. This is why the mean is lower compared to the previous value.
Upon conducting various analyses, it is evident that when the use of virtual humans is disclosed, people’s perception tends to shift towards the negative. This shift is particularly noticeable in terms of the likelihood of their usage and overall acceptance. The correlation analysis in
Table 12 shows whether the factors impact each other and whether they have any statistically significant relationship.
This analysis shows that trust had a strong positive correlation with usage likelihood before the reveal, at 0.851 and 0.767. After the revelation, trust still had a very strong positive correlation with usage likelihood, while performance was not observed to have as strong a relationship with the other variables. We conducted a regression analysis using pre-disclosure data (
Table 13), with overall acceptance as the dependent variable and the other three factors—trust (trb), performance (pb), and usage likelihood (ub)—as independent variables.
Here, we observe that all three factors (trust, performance, and usage likelihood) have a p-value of <0.001, indicating statistical significance. However, these are unstandardized coefficients, meaning their magnitudes cannot be directly compared, as they are expressed in different units. Thus, while all three factors significantly influence overall acceptance, we cannot infer that they do so equally based on their coefficients alone. Still, trust is said to have a significant relationship with just overall acceptance, as the regression with trust as the dependent variable showed that its p-value is above 0.05 for performance and usage likelihood. When we conducted a regression test using the post-revelation values, we discovered that overall acceptance is influenced by trust and the likelihood of usage. At the same time, performance does not seem to have a significant impact, with a p-value of 0.099.
To explore the factors that influence trust specifically after the VH disclosure, we conducted a second regression analysis (
Table 14) with trust as the dependent variable. This analysis allows us to investigate whether performance, usage likelihood, and overall acceptance contribute to changes in trust following the disclosure of a virtual human’s nature. We see after the revelation that it has a statistically significant relationship with usage likelihood and overall acceptance, while performance has no impact on trust. The regression analysis indicates that the VH’s pre-disclosure overall acceptance was significantly influenced by trust (β = 0.327,
p < 0.001) and usage likelihood (β = 0.391,
p < 0.001), suggesting these factors are critical determinants of acceptance. Post-disclosure, trust remains a significant predictor (β = 0.59,
p < 0.001), though the influence of performance diminishes, reflecting the impact of the erosion of trust after the disclosure. Overall, this study’s findings show that trust is a critical factor that can influence the adoption of virtual humans, and the general perception towards the technology is found to be mostly negative outside of the social media context.
The thematic analysis revealed a dichotomy in participants’ responses, with some viewing the technology as innovative and exciting, particularly in social media contexts, while others expressed significant concerns about deception and the ethical implications, especially in business and personal settings. During the qualitative analysis, the two open-ended questions and 30 interviews were analysed. The first question focused on understanding what led to their change in perception regarding virtual humans following the revelation. Across the experiments, the themes identified could be classified or grouped as positive or negative reactions. In the case of Experiment 1, within the social media context, people report the technology as being exciting, and some even mentioned that their positive attitude in the experiment after the revelation was because initially it felt boring, but after the revelation, it felt more futuristic. Not only positive reactions were provided in the comments and interviews. Many felt betrayed, concerned, and worried about security, which shows that some had a positive shift in perception while others had a negative shift in perception. As for Experiment 2, some had a positive attitude, mentioning that this is better than the existing chatbots, showcasing that using the VHs is an improvement. At the same time, their concerns about the technology’s potential applications were growing, and their unfavourable opinions were far more prevalent. These concerns centred on how AI could exploit their data or interactions. They also point out that businesses may influence people’s opinions by seeming to be human instead of acting honestly. Although the open comments in Experiment 3 and in the interviews continue to emphasise the positive aspects of the technology and its superiority over the current system, it was determined that the negative responses—such as security concerns, a lack of human touch, and a sense of betrayal—were more pertinent. In addition, the respondents stated that even if it were possible, they would not utilise VHs. This demonstrates the unfavourable change in perspective following the disclosures, which was shown to be more substantial in Experiments 2 and 3. This shift was caused by people’s anxieties and feelings about the technology, which only some individuals continued to trust or think was beneficial.
The second open-ended question asked about people’s opinions of using virtual humans in various experimental contexts. The results showed that while most people responded positively to using VHs in social media and thought they were amazing and exciting, some people were sceptical about how they could affect people and aid in spreading false information. In the second experiment, there were no completely positive reactions, with most being negative or mixed. The respondents and interviewees highlighted their fear of the lack of transparency and the loss of human touch that the adoption would create. The participants also mentioned that while they are happy that chatbots are improving, it does lead to a more impersonal approach from companies that also leads to questions about the quality of the service provided. In Experiment 3, most comments focused on ethical concerns and people’s preference for human interaction over virtual humans who do not understand them. In contrast to the other two experiments, where people were generally willing to accept virtual humans, many people were sceptical of VHs and their role in spreading misinformation on social media. These qualitative insights help explain the quantitative findings, where trust and overall acceptance decreased most sharply in business contexts, likely due to concerns over transparency and data security.
This study’s findings are consistent with previous studies, which highlighted the positive impact of virtual influencers on consumer purchase intentions driven by trust and emotional attachment (
Jiménez-Castillo and Sánchez-Fernández 2019;
Awdziej et al. 2020;
Gerlich 2023b;
De Cicco et al. 2024). The results from the social media experiment revealed that even after participants were informed about the use of virtual humans, their trust in the scenario increased, and their overall acceptance of the technology was better than before they knew it involved virtual humans. The regression analysis found that trust is one of the main factors influencing the overall acceptance of technology, for better or worse. There are studies of business and work settings where VHs carry out some of the work processes that could help or impact peer productivity (
Hill 2014;
Gürerk et al. 2018).
Lee et al. (
2015) mentioned that integration is only possible after a careful evaluation of acceptance, and this seems to be the major finding of this study, as in a business context, people do not trust or accept the technology as much as within social media. There is a growing reluctance due to fear of security breaches and a decrease in personal interactions. This makes the service less friendly and of a lower quality, as people feel unimportant to the company. In general, people were more opposed to the adoption of virtual humans in technology. They emphasised the importance of transparency regarding the use of technology, as concealing the use of artificial intelligence was often seen as a betrayal of the people involved in the experiment. People are concerned that this technology could be misused and lead to more problems; therefore, it is not an ideal candidate to replace humans. The absence of personal interaction further worsens relationships. The survey participants believe that although improvements in areas such as chatbots are possible, these should be restricted to specific changes, and companies should not attempt to replace the current system without implementing transparent protocols and ethical guidelines.
Overall, this study was able to show that in contexts like social media or digital marketing, companies find increasing uses for VHs. People accept them to a certain extent, but the adoption of VHs in the field cannot extend beyond the existing comfort level of the people. Their adoption in business might be challenging as people do not trust the technology even in general settings. Creating guidelines and security protocols regarding the use of VHs could increase confidence among the people. Still, other elements like attitudes and feelings of importance cannot be addressed, and marketing strategies need to explore new ways in which VHs could be used in creating this feeling among people. Many areas or factors need to be tackled in the long run for their widespread adoption, and this study helps identify the current perception of the technology. It helps us understand where improvements can be made.
5. Conclusions
This study provides a comprehensive examination of how virtual humans (VHs) are perceived across different contexts—namely, social media, business environments, and general public interactions. The findings underscore the complexity and variability of these perceptions, revealing significant differences in trust, performance, usage likelihood, and overall acceptance before and after the disclosure of the VH’s true nature.
5.1. Summary of Key Findings
In the social media context (Experiment 1), this study found that trust levels, while initially positive, experienced a slight decline post-disclosure, moving from an average of 2.1487 to 1.9795. This suggests that while social media users may generally have a favourable view of VHs, there remains a degree of scepticism once their artificial nature is revealed. Their performance perception in this context showed a marginal increase from 1.3538 to 1.4712, indicating that the technical execution of VHs in social media might be appreciated even more after their nature is disclosed. However, their usage likelihood decreased from 2.5846 to 1.9077, suggesting that while users might appreciate the novelty and technical prowess of VHs, they may hesitate to engage with them frequently after learning about their artificiality. Their overall acceptance followed a similar trend, slightly decreasing from 2.1526 to 1.9833, which reflects a positive but cautious attitude.
In the business/professional environment (Experiment 2), an interesting divergence from the expected outcome was observed. The trust in VHs increased from 1.1290 to 1.3044 after disclosure, suggesting that transparency in professional settings can actually enhance trust. This indicates that business users might value honesty and openness about the use of technology, which could mitigate some of the negative impacts of discovering the non-human nature of their interlocutor. The performance ratings of VHs also saw a slight improvement from 1.8360 to 1.9839, remaining in the extremely positive range, which further supports the notion that performance is less impacted by the revelation of artificiality in professional contexts. Their usage likelihood showed a small increase from 1.4194 to 1.8871, reflecting a continued willingness to engage with VHs in a business setting, as the slight rise still indicates a generally positive outlook. The overall acceptance of VHs similarly increased, though modestly, from 1.7961 to 2.0737, suggesting that business users remain open to the integration of VHs, provided that their capabilities are clear and their use is transparent.
Conversely, the general public perception (Experiment 3) demonstrated the most significant decline in all metrics following the VH disclosure. Trust plummeted from 1.0684 to 2.7179, indicating a sharp drop from extremely positive to a more neutral or even slightly negative perception. This suggests that the general public is particularly sensitive to the artificiality of VHs and may react strongly to their use when their nature is revealed. Performance also decreased, though less dramatically, from 1.0983 to 1.6937, indicating that while the technical performance of VHs was still viewed positively, it was less appreciated after the disclosure. Their usage likelihood experienced the most drastic change, skyrocketing from an extremely positive 1.0812 to a highly negative 4.6496. This sharp shift highlights a severe reluctance to engage with VHs post-disclosure, likely driven by discomfort or mistrust. Similarly, their overall acceptance fell from 1.1154 to 2.3533, showing a clear move from a very positive to a more cautious and critical stance. While the study reveals statistically significant shifts in trust, performance, usage likelihood, and overall acceptance following the disclosure of virtual humans, the effect sizes (Cohen’s d) highlight the practical implications of these changes, particularly in public settings, where the trust in virtual humans was most impacted.
5.2. Interpretation and Broader Implications
The results of this study have significant implications for the deployment and acceptance of virtual humans across different sectors. The slight but consistent drop in trust and acceptance in the social media context suggests that while users may be initially intrigued by VHs, their long-term engagement may be jeopardised if their artificial nature is disclosed without careful consideration of user expectations and transparency. In business environments, the data imply that transparency about the use of VHs can enhance trust and acceptance, suggesting that corporate communication strategies should emphasise honesty and clarity regarding the use of digital agents. However, the sharp decline observed in the general public context raises concerns about the broader societal acceptance of VHs, particularly in settings where personal and ethical considerations are paramount.
5.3. Challenges and Ethical Considerations
This study also sheds light on the ethical challenges associated with the use of VHs. Participants expressed concerns about data privacy, the potential misuse of information, and the erosion of human relationships, particularly in professional settings where the reliance on digital agents might weaken the perceived value of human interactions. These concerns are compounded by the lack of established ethical norms and regulations governing the use of VHs, particularly in contexts where their use is not immediately apparent to the user. The fear of transitioning to a machine-dominated system that could diminish the importance of human agency is a significant barrier to the widespread adoption of VHs. Therefore, addressing these ethical concerns is crucial for fostering trust and ensuring that VHs are integrated in a way that complements, rather than replaces, human interaction.
5.4. Study Limitations
While this study provides valuable insights into the perception of VHs across different contexts, several limitations must be acknowledged. First, the sample size, while adequate for the study’s scope, may not fully capture the diversity of the opinions and attitudes present in the broader population. The relatively small sample size could limit the generalizability of these findings, particularly when considering the vast range of cultural, social, and technological contexts in which VHs may be deployed. Additionally, the study was geographically concentrated, primarily involving participants from a specific region. This geographic limitation may have influenced the results, as cultural and regional factors often play a significant role in shaping perceptions of technology.
Another limitation concerns the artificial nature of the experimental settings. While the simulations were designed to mimic real-world interactions, they cannot fully replicate the complexity and nuances of actual social, business, or public environments. Participants may have responded differently in more naturalistic settings, where the stakes and social dynamics are more pronounced. Furthermore, the study relied on self-reported data, which, while useful, can be subject to biases such as social desirability or recall bias. These limitations suggest that while these findings are indicative, they should be interpreted with caution and complemented by further research.
Although the study included participants from three distinct age groups, no significant differences were found in their responses. This suggests a relatively uniform perception of virtual humans across the age spectrum studied. However, future research could explore whether more nuanced age-related differences emerge in larger or more diverse samples, particularly when comparing younger versus older adults in different cultural or professional contexts.
This study did not specifically analyse gender or cultural influences, which are known to affect AI perceptions. Future studies should include a gender-based analysis and consider cultural differences, particularly as AI technologies become more prevalent in global markets. Understanding these variables could lead to more tailored and effective AI interactions.
Although this study did not collect data on education levels, this factor could influence the acceptance of AI. Participants with higher levels of education might have a better understanding of AI technologies, potentially leading to greater trust and acceptance. Future research should consider including education as a variable to explore its impact on the perception of virtual humans.
5.5. Future Research Directions
The findings of this study open up several avenues for future research. Expanding the sample size and exploring cross-cultural differences in the perception of VHs would provide valuable insights into how cultural and regional factors influence acceptance and trust. Additionally, investigating the long-term effects of repeated interactions with VHs could help determine whether initial scepticism can be mitigated over time through familiarity and improved technological integration. Further research should also explore the development of ethical guidelines and regulatory frameworks that address the unique challenges posed by VHs, ensuring that their use is both transparent and aligned with societal values.
5.6. Practical Implications
For businesses and developers, this study’s findings underscore the importance of maintaining a human touch in interactions involving VHs. Ensuring that users are fully informed about the nature of their interactions and that ethical standards are upheld will be key to fostering long-term relationships and trust. Policymakers should consider the implications of these findings for the development of regulations that govern the use of VHs, particularly in contexts where the technology has the potential to impact social and professional relationships significantly.
While virtual humans present exciting opportunities for enhancing digital interactions, their successful integration into various contexts will depend on carefully navigating the challenges of trust, ethical considerations, and user acceptance. This study highlights the importance of context in shaping perceptions of VHs and suggests that transparency and ethical practises will be essential in ensuring that this technology is embraced by the broader public. Future research and practical applications should focus on addressing the concerns identified in this study to promote the development of a more positive and widespread acceptance of virtual humans in both professional and everyday settings.