1. Introduction
The rapid increase in the number of older adults (75 years old and up) is a major concern of the developed countries [
1]. In the 1950s, the probability that an 80-year-old would survive to age 90 was 15–16% for women and 12% for men. By 2002, this percentage had increased to 37% and 25%, respectively. Since 1840, female life expectancy has increased by approximately two years per decade, worldwide. This linear growth has yet to stagnate and therefore implies that ‘the human life span is not closing in on its limit’ [
2]. Life expectancy at age 20 is predicted to increase by approximately one year per decade for females and males between now and 2040 [
3]. Compared to the working-age population, the proportion of older adults will eventually grow to a point where there are not enough people, let alone professional care-givers, to care for the elderly [
4,
5,
6]. It seems impossible to continue organizing eldercare in the same way as before [
7]. Thus, the search for innovative solutions has become crucial, and the deployment of robots in healthcare seems to be a promising way forward [
8,
9,
10,
11,
12,
13]; however, robot care is not without some controversy.
Care-robotics can be divided into three types: companion, assisting, and monitoring robots [
6]. Assisting robots are used to help patients with their daily tasks (e.g., washing and dressing) and monitoring robots are used to keep an eye on the patient (e.g., medicine intake, fall detection). Companion robots are designed to counter social isolation and loneliness. Most people, however, feel uncomfortable with such robots [
14], and if care professionals are hesitant from the start, a possible game changer may be lost without having been properly tried and tested.
A previous study elaborated on the effects of education on moral considerations regarding care technology in a quantitative examination among higher and lower educated care professionals [
15]. This previous study found that the acceptance of robots in care was more strongly associated with the participants’ moral considerations than with utility. Hence, the current study aimed to further explore the educational context and conduct research within the group of vocational educated students. We wanted to collect and examine the opinions of students of care and welfare programmes of Dutch Vocational Educational Institutes, because they are the care professionals of the future. Students in healthcare will form a large part of the group of care professionals that, in the future, will work with this kind of technology [
16].
Whereas the previous study [
15] explored the possible differences between intermediate and higher vocational educated students in care, this current study will examine educational differences
within the group of intermediate vocational educated students. In the Netherlands, intermediate Vocational Education and Training (VET) is called Middelbaar Beroeps Onderwijs (MBO). VET is the main supplier to the labour market and is often regarded as the ‘foundation of the economy’ or the ‘backbone of society’. Approximately 40% of the Dutch working population has completed a vocational curriculum to at least an intermediate vocational training level [
17]. Intermediate vocational training is divided into four levels in the Netherlands. VET offers programmes on four different levels, ranging from the entry level to the middle-management level. Thus, it is interesting to study whether or not educational differences within the different levels of VET lead to different appraisals of care robots.
Therefore, in the current study, we attempted to quantify the various opinions of Dutch vocational care students on care robots from the perspective of the four principles of biomedical ethics [
18]: beneficence, non-maleficence, justice, and autonomy. From a technology acceptance perspective, we added ‘expected utility of the robot’ and ‘intentions to use the robot’ to this list. Since age, skillfulness, and interest in technology also contribute to opinion formation [
19,
20,
21] we studied these variables as well for their possible moderating effects.
Beauchamp and Childress [
18] have proposed a system of moral principles in the practice of medicine. Maleficence is derived from the principle of “non-maleficence”, which states that technology should “first, do no harm” (
primum non nocere). For many, this is the paramount principle of biomedical ethics. Beneficence holds that technology should promote the well-being of the patient. Justice is a concern for fairness and equality, while autonomy is the patient’s right and ability to freely make decisions about medical treatment. In this study, we aimed to identify which of these principles dominated the moral considerations regarding robots among trainee care professionals. Our research question, then, is:
How do care and welfare students at diverse levels of vocational education differentially evaluate healthcare robotics in relation to their daily routines?
We used the following sub-questions:
RQ-1: Do prospective care professionals perceive different types of robots (Assistive, Monitoring, and Companion) differently in terms of occupational ethics (Beneficence, Maleficence, Justice, Autonomy)?
RQ-2: Do perceptions of healthcare robots differ between vocational students at lower and middle levels? Related, do they perceive the various robot types differently?
RQ-3: How do prospective care professionals evaluate care robots in terms of utility and possible use intentions? Related, how do these evaluations differ per robot type?
3. Data Collection
Participants received a link to an online questionnaire that was specifically developed for this study. The questionnaire had three versions, one for assistive, one for monitoring, and one for companion robots; each version consisted of 39 questionnaire items. The items per robot type were identical so they could be compared. Demographics were probed using seven additional questions.
After opening the link in Qualtrics (Version 24(892), Provo, UT, USA)a brief introduction and consent form was presented. Upon agreement, the participant could commence the questionnaire. Participants were randomly assigned to one of the three versions. The questionnaire opened with a picture of a healthcare robot (
Figure 1) followed by a brief description of its capabilities and the tasks it could execute (e.g., reading aloud, exercise coaching, or reminding of medicine intake).
4. Measures
We measured six theoretical variables, which as a group we will term appraisal domains (i.e., perceived beneficence, maleficence, justice, autonomy, utility, and use intentions), four educational levels (levels 1–4), and three robot types (assistive, monitoring, and companion) along with six control variables: gender, age, school year, technical skills, interest in technology, and enjoying working with computers. Education Level was an ordinal variable: lower vocational education (levels 1 and 2) and middle vocational education (levels 3 and 4). The theoretical variables were measured at quasi interval level with initially six Likert-type items, each rated on a six-point scale (1 = strongly disagree, 2 = disagree, 3 = slightly disagree, 4 = slightly agree, 5 = agree, and 6 = strongly agree). Principal Component Analyses and Reliability analyses revealed that several items had to be discarded to form reliable scales. The questionnaire items that were included per appraisal domain can be found in
Appendix A. Scale reliabilities were calculated with Cronbach’s alpha as reported in
Table 1, using 0.7 as the cut-off point to decide whether a scale was sufficiently reliable or not [
26]. The Justice scale failed the 0.7 criterion and had to be discarded entirely. Apparently, the way we measured justice did not converge into a solid underlying concept. All other scales performed well (
Table 1). After recoding the counter-indicative items, the items on each scale could be summed and averaged to calculate a mean index (
Table 1).
Although 2365 participants completed the questionnaire, cells were not filled equally. Unfortunately, only 2 participants studied at level 1, which made it necessary to drop that level from our analyses. Then, we had to exclude another 38 participants because they had completed the questionnaire in an unreasonably short period of time (which was recorded by Qualtrics) and/or because they checked the same rating scale answer for each item. That left us with
N = 2325 participants in the final analysis. Their characteristics can be found in
Table 1.
Table 2 provides the means and standard deviations of the appraisal domains.
Table 1 shows that participants were evenly distributed over the three robot types (33% each).
Table 1 also shows an overrepresentation of females, although in the care professions that is a valid ecological outcome [
27]. Unfortunately, however, over 50% of the total sample studied at vocational level 4 and a mere 9.1% at level 2, which seriously jeopardized our questions on differences in perception between educational levels. Almost the entire sample self-reported that they were skilled computer users.
5. Results
We calculated the grand mean scores of our measurement scales (
Table 2) and ran a GLM Repeated Measures for the 5-leveled within-subjects factor of appraisal domain (beneficence vs. maleficence vs. autonomy vs. utility vs. use intentions) by the between-subjects factors of robot type (assistive vs. monitoring vs. companion) and education level (2 vs. 3 vs. 4).
Rating of Appraisal Domains through Ethics
The interaction between robot type, education level, and appraisal domain was not significant (p = 0.296) nor was the interaction between robot type and education level (p = 0.672).
However, the main effect of appraisal domain was significant with an intermediate effect size (Wilks’ λ = 0.53, F(4,2313) = 518.67, p = 0.000, ηp2 = 0.473), indicating that independent of robot type or education level, participants scored lowest on maleficence (M = 2.81, SD = 0.95), then use intentions (M = 3.08, SD = 1.19), then beneficence (M = 3.64, SD = 1.13), higher on autonomy (M = 3.78, SD = 0.81), and highest on utility (M = 3.84, SD = 0.93).
Rating of Appraisal Domains by Robot Type
The interaction between robot type and appraisal domain was significant (
λ = 0.99,
F(8,4626) = 4.00,
p = 0.00,
ηp2 = 0.007), which was supported by a significant main effect of Robot Type (
F(2,2316) = 5.23,
p = 0.005,
ηp2 = 0.004). To further scrutinize the interaction effect, we ran a number of independent samples
t-tests (two-tailed) with Bonferroni correction (
α = 0.05/15 contrasts = 0.003) [
28], indicating that independent of education level, participants deemed monitoring robots more beneficent than assistive robots (
t(1550) = 3.77,
p = 0.000) and assistive robots more beneficent than companion ones (
t(1549) = 3.59,
p = 0.000).
Assistive robots were judged as more maleficent than monitoring robots (t(1549) = 4.24, p = 0.000) and more maleficent than companion robots (t(1545) = 4.56, p = 0.000). Monitoring robots were perceived as having more utility than companion robots (t(1550) = 5.01, p = 0.000) and as a trend, assistive robots were also judged as having more utility than companion machines (t(1545) = 2.79, p = 0.005).
Participants indicated higher intentions to use monitoring than companion robots (t(1550) = 3.46, p = 0.001). All other comparisons were not significant.
Rating of Appraisal Domains by Educational Levels
The interaction between education level and appraisal domain was significant as well (λ = 0.99, F(8,4626) = 4.47, p = 0.000, ηp2 = 0.008). This interaction was supported by a significant main effect of education level (F(2,2316) = 4.69, p = 0.009, ηp2 = 0.004). We then ran a number of independent samples t-tests (2-tailed) with Bonferroni correction (α = 0.05/15 contrasts = 0.003), indicating that independent of robot type, education level 4 students perceived more beneficence than level 3 students (t(2111) = 2.94, p = 0.003), whereas level 3 saw, as a trend, more maleficence than level 4 (t(2111) = 2.72, p = 0.007). Level 4 also assigned more utility to robots than level 3 students (t(2111) = 4.47, p = 0.000), and as a trend, level 4 thought more than level 2 that robots increased the autonomy of patients (t(1417) = 2.81, p = 0.005). All other comparisons were not significant.
Overall Ratings of Appraisal Domains
In sum, independent of robot type or education level, maleficence and use intentions scored lowest while autonomy and utility scored highest. Independent of education level, participants judged that monitoring robots were more beneficent than assistive robots and assistive robots more beneficent than companion robots. Assistive robots were perceived as more maleficent than monitoring robots and more than companion robots. Monitoring robots had more utility than companion robots and as a trend, assistive robots also had more utility than companion machines. Participants indicated a firmer intention to use monitoring than companion robots. All other comparisons were not significant.
With respect to education level, independent of robot type, level 4 students perceived more beneficence than level 3 students, whereas level 3 saw, as a trend, more maleficence than level 4. Level 4 also assigned more utility to robots than level 3 students, and as a trend, level 4 deemed that robots increased the autonomy of patients, more so than level 2.
6. Conclusions
The purpose of the current study was to determine which of the four principles of Beauchamp and Childress [
18] were most prominent in the estimations of lower and middle vocational care students with regard to working with robots in their future care practice.
Overall, students scored Maleficence the lowest, meaning that care robots in general were not seen as pernicious. However, students of care also rated use intentions low, indicating poor motivation to actually use a robot in the future. Possible beneficent effects of care robots were received with relative neutrality, whereas the potential increase in a client’s autonomy was deemed considerable. Highest scores were obtained for the utility of robots, which is surprising in view of the students’ reluctance to use them. On the whole, level 4 students were slightly more positive than level 3 or 2 students with respect to beneficence, utility, and the client’s autonomy. All students viewed companion robots as more beneficent than monitoring robots and monitoring robots more beneficent than assistive robots. Although regarded as more maleficent, monitoring and assisting machines were also seen as more useful.
In all, these care students saw little harm in robots, found them useful, and expected clients to become more independent because of them; on the other hand, they were quite hesitant about using robots in their future practice. It is possible that the perceived potential effectiveness of robots for their work practice was affected by fear of job loss.
7. Discussion
One might think that current students of care have already acquired so-called ‘21st century skills’, that is, they would have “information literacy, media literacy, and information, communication and technology literacy” [
29]. Additionally, if that is the case, so theory has it, perceived usefulness and a positive attitude toward technology should increase the intention to use technology. Indeed, computer self-efficacy can act as an antecedent for perceived usefulness and positive attitude towards computer use [
30]. 93.6% of our survey participants stated they were skilled computer users and 72.3% claimed to enjoy working with computers. Nevertheless, these 21st-century skilled students were not very eager to employ robots in their future work practice.
Admittedly, the vocational care students in this survey did not see much harm in care robots, for the patient that is. However, when it came to their future work practice, they were reluctant to envision employing care machines, in spite of their potential utility. Listed hereafter are a few possible explanations for this.
Possible Explanations for the Hesitation in Using Care Robots
Robots may not have been viewed as maleficent, but neither were they seen as being beneficial or helpful (beneficence was rated ‘neutral’). Although useful for the instrumental side of nursing and caring (utility high), it might be that these students feared that robots would weaken the relationship between caregiver and care receiver (autonomy high), commonly considered fundamental to ‘good care.’ Hence, they were hesitant to work with robots. Lewis and West [
31] state that the care relationship is crucial to securing care quality. If ‘good care’ depends substantially on the quality of the care relationship, then more attention should be paid to the human care workforce. Perhaps this was a principal concern of our sample of care students. Moreover, in covering the instrumental aspects of care, robots may also have been perceived as an occupational threat, making the nurse seem less important (e.g., when the client becomes attached to the robot) or even redundant.
On a societal plane, 70% of Europeans have a positive attitude towards robots in general [
32]. However, when it comes to healthcare robots, they are not so positive. Only 22% of Europeans think that robots should be introduced in healthcare [
32]. Sixty percent are even resentful of the prospect of robots caring for children and older adults [
32]. Perhaps our care students were resonating with a more public trend, vented in the media and discussed at the coffee table.
We did not find great differences between educational levels in the way they perceived care robots. It could be that the educational differences between lower and middle vocational care students are negligible. As an additional exploratory analysis, we combined levels 2 and 3 and compared them with level 4, performing chi-square analysis is (
Table 3). We found that level 4 students perceived more beneficence in care robots. However, contrary to our expectations, they do not perceive less maleficence. Higher vocational students also perceived a greater utility in care robots than lower vocational students, however this higher perceived utility does not translate into a higher use intention. All care students expressed low use intentions, the potential utility of robots notwithstanding.
As advised by Holloway and Wheeler [
33], it may be more worthwhile to turn to qualitative research approaches to resolve matters of change or conflict, particularly in care relations, and explore the behaviours, feelings, and experiences of care students in confrontation with robots on the work floor in more detail.
Trainee care professionals perceived particular types of robots differently in terms of occupational ethics and use. Our research challenge is to ascertain what these care professionals would want from an assistive or monitoring machine that is—in their eyes—potentially very useful and somewhat innocuous. Should a robot assistant have a moral reasoning system that tells it what is ‘good care’? Should a monitoring device know what information is private and what should be disclosed to the nurses at the ward? Companion robots may be seen as beneficent, but how can they become more useful? Additionally, how should they behave such that they do not undermine the quality of the care relationship between humans? These are not questions of occupational ethics and utility alone, as their answers will encourage better partnerships between human caretakers and artificial systems of the future. More research is needed, with a more rigorous research methodology to truly obtain the objections raised by (trainee) caregivers to facilitate acceptability. It is stated in the literature [
34] that healthcare robots can potentially enhance elderly well-being and decrease the workload on caregivers.
Education could Make a Difference
This study suggests that it would be wise to enrich the curriculum of intermediate care education with practical classes on the use and ethical implications of new care technology, particularly in the case of care robots. This is especially of interest in the intermediate vocational domain because this group will be the main supplier of direct care of older adults in the near future. It is important to ensure that the provided care is in line with the wishes and needs of our vulnerable older adults, but at the same time keeps the job satisfaction of our precious caregivers as high as possible, given the circumstances.