Next Article in Journal
On Being Twice Exceptional in Sweden—An Interview-Based Case Study about the Educational Situation for a Gifted Student Diagnosed with ADHD
Previous Article in Journal
The Role of Relationships in Resilience: Teachers Who Were At-Risk Youth Supporting At-Risk Students
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning and Motivation When Using Multiple-Try in a Digital Game for Primary Students in Chile

by
Claudio Cubillos
1,*,
Silvana Roncagliolo
1 and
Daniel Cabrera-Paniagua
2,*
1
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2362807, Chile
2
Escuela de Ingeniería Informática, Universidad de Valparaíso, Valparaíso 2362905, Chile
*
Authors to whom correspondence should be addressed.
Educ. Sci. 2023, 13(11), 1119; https://doi.org/10.3390/educsci13111119
Submission received: 22 August 2023 / Revised: 24 October 2023 / Accepted: 3 November 2023 / Published: 8 November 2023
(This article belongs to the Section Technology Enhanced Education)

Abstract

:
The number of attempts to provide students is a key instructional characteristic in computer-based learning (CBL). However, it has not been covered extensively, and there is a need to delve deeper into the factors affecting multiple-try performance and allowing its successful use, including the learner’s involved emotional processes. This study examines the effects of multiple-try on a drill-and-practice mathematical game devoted to primary school students. A total of 73 students from four courses from two schools participated in the experiment. They were randomly assigned to a 3-attempt multiple-try (MTF) and a single-try knowledge of correct response (KCR) conditions. The study covered impacts on learning performance, together with motivation, effort, pressure, and the value of students regarding the learning activity based on the self-determination theory (SDT) perspective and its cognitive evaluation sub-theory (CET). The study’s main findings were that (a) the MTF condition outperformed KCR in terms of students’ learning gains; (b) MTF presented higher levels of perceived competence and autonomy, which, according to SDT, fosters motivation and learning; (c) a cost was yielded in students’ perceived pressure under MTF; and (d) perceived effort and value was similarly high for both conditions despite learning differences. This study complements the existing literature on multiple-try, providing insights into what conditions are beneficial for multiple-try use.

1. Introduction

Computer-based learning (CBL) has been raised over the years as an exciting complement to traditional education [1,2,3].
The success of digital educational training depends on various factors associated with the activity, such as instructional strategies and task complexity. Still, other factors are associated with the student, such as prior knowledge, metacognitive knowledge, and motivation [4]. Within the instructional characteristics, we find the number of attempts offered to the student to answer each question. Many studies have assumed a single try per item, although others have considered the possibility of providing multiple try to answer the same question [4].
Multiple-try has its roots in the pioneering work of Pressey [5,6] involving a technological device for automated testing. It provided immediate feedback on the result (KR), allowing the learner to try again until the right response was reached, thus coining the expression of answer-until-correct (AUC).
Nowadays, there is a greater possibility of using multiple-try mechanics in learning management systems like Moodle and H5P, and online educational games like Educaplay.
Despite its early beginnings and possibilities of actual use, diverse authors agree that the existing literature considering multiple-try is scarce [7,8], especially under a digital formative context. Currently, using multiple attempts by practitioners and researchers is not easy because the conditions for its effective use are not clear.
This problem may be partially explained because this field traditionally considers multiple-try as a feedback-type feature [9]. Hence, many past meta-analyses considered multiple attempts as part of elaborated feedback or knowledge of results [10,11,12,13]. More recently, [4] and [14] have classified multiple-try as a separate feedback type. But still, few reviews exist tackling multiple-try and its effects, besides being outdated [15,16,17]. Therefore, it becomes difficult to measure the impacts of multiple attempts across conditions and provide straightforward conclusions about multiple-try effects.
Another issue is that past research regarding multiple attempts has mainly focused on students’ learning effects rather than searching for the conditions under which the use of multiple attempts is effective, together with the circumstances that allow such positive effects to be long-lasting and efficient [18].
While some studies report substantial learning gains when using multiple-try [8,16,19], others describe no significant differences when contrasting it to single-try solutions [20,21,22].
In particular, the review by Van der Kleij et al. [17] considered 18 studies, where six involved try-again solutions. None of them obtained significant positive effects for multiple-try when considering diverse timing, feedback level, and task complexity conditions except for Murphy’s work [23]. His delayed elaborated feedback with hints + try again performed significantly higher with students working in pairs but not when working individually.
The existing literature on instructional feedback points out that diverse situational variables influence student learning and performance in educational contexts [4,18]. Research exists analyzing the moderating role of task complexity [13,14] and previous knowledge [24] in student learning outcomes. However, in what contexts multiple-try can be administered effectively and successfully is still open.
We argue that an adequate control on these two variables, task complexity, and prior knowledge, should promote positive effects on multiple-try use. There are some insights on MTF being preferable for high-order learning outcomes, especially in text-comprehension scenarios [13,16] and vocabulary lessons [25]. However, there is no consensus on whether those findings extend to the K-12 math domain or what levels of prior knowledge are appropriate for MTF use.
Another aspect is the role that motivation and other emotional variables have in the learning process, either under traditional education [26,27] or in computer-based learning (CBL) settings [28]. More focus is needed “on the processes induced by feedback interventions and not on the general question of whether FIs improve performance” [18]. Hence, further attention is needed on what happens to students internally and emotionally when using multiple tries.
Therefore, the present study objective is to explore the impacts of using multiple-try feedback (MTF) and knowledge of correct response (KCR) feedback under a drill-and-practice mathematical digital game devoted to primary students. It focuses on three high-order math learning objectives: 3-digit number identification, ordering, and money counting.
Our study also uses self-determination theory (SDT) and its cognitive evaluation sub-theory (CET) [29,30] to investigate the emotional factors that are affected by feedback and, in particular, by multiple-try use. SDT is a model of human motivation that has been revealed as appropriate, as it has been used to explain students’ emotional processes in other contexts of computer-based learning [31,32]. SDT postulates that the psychological needs for competence, autonomy, and relatedness must be fulfilled to foster motivation.
Competence refers to the need “for being effective in one’s interactions with the environment” [29], and autonomy involves experiencing “freedom in initiating one’s behavior” [33]. Autonomy and competence are necessary for a good learning environment with greater motivation and creativity [29].
The CET theory complements the above, holding that high levels of pressure negatively affect motivation, while adequate levels of effort are indicators of a suitable level of challenge and the absence of controlling situations that may negatively affect motivation [30]. Finally, there is the level of the learner’s perceived value, indicating the degree to which the behavior is internalized and integrated into his or her personal values and the degree of self-determination [33].
No previous works have been found that use SDT or CET as a theoretical model to explain the effects of multiple-try. Therefore, the main contributions to the literature in this field are as follows:
  • To develop a successful experience using multiple tries under school mathematics and its effects on students’ learning, motivation, and related emotional processes.
  • To explain why multiple attempts generate certain learning effects from the SDT + CET perspective, mainly by developing a greater sense of competence and autonomy in students while fostering suitable levels of effort and pressure.
  • Progress in devealing under what conditions, primarily related to prior knowledge and in a K-12 math setting, the use of multiple-try generates positive results regarding learning, motivation, and self-determination.

1.1. Feedback and Attempts

Feedback can be understood as providing the student with some information about correcting an answer [34] and “Can help students identify and correct errors and misconceptions, develop more effective and efficient problem-solving strategies, and improve their self-regulation.” [13]. From a more general perspective, feedback corresponds to any piece of information given by an agent regarding a person’s performance or understanding, being such an agent as a teacher, peer, book, parent, and experience [35,36].
The scope of feedback in the present study focuses on the task level and its learning (formative) instead of summary feedback [14] or feedback focusing on task motivation or the self [18].
Most research has focused on the type of information provided in the feedback. Starting from merely informing students if the answer was correct or not [37], named knowledge of result (KR), and followed by specifying the right answer without further explanations, called knowledge of correct response (KCR) [38,39]. Finishing with more elaborate feedback (EF) presented as explanations, strategies, hints, informative tutoring [13], or interactive feedback [35] to arrive at the correct answer.
Besides these types based on message content, we can also consider the number of attempts the student will have upon an incorrect response before moving to the next exercise or question. It can also be seen as the number of learner expositions to the same item before moving on.
Regarding implementations of multiple attempts, in between single-try and the early answer-until-correct, multiple-try feedback (MTF) exists. In its simplest form, MTF gives students a limited number of attempts and KR between tries [4]. The literature presents different ways of implementing multiple-try by varying the number of trials given (2, 3, 5, and until-correct) and the feedback information provided between tries (KR, KCR, or EF). These diverse implementations make it hard to provide any generalizable conclusions about which multiple-try implementations are preferable under what situations.
A close relationship exists between the number of attempts and the feedback type delivered to the student. Both are part of the instructional characteristics to be delivered within the learning process, among other aspects such as the item type. In this sense, multiple-try was often considered a type of elaborated feedback because it performs a directive function in informing about the answer’s correctness and giving another opportunity upon error. However, aside from such a directive function, MTF provides an interactive mechanism [4] with multiple immediate exposures [40] to the same question. These features are not present in any other type of elaborated feedback, so it is better to consider the number of attempts in a separate dimension.
MTF has both immediate and delayed feedback features. It is similar to single-try (STF) regarding immediacy concerning feedback timing. However, it is also similar to delayed feedback (DF) regarding the number of direct exposures to the problem or question. In this sense, Clariana, Wagner, and Murphy [40] observed that MTF provided learning effects midway between single-try and delayed feedback.
Different theories have been used to explain the effects of multiple-try. Many authors agree that for feedback to affect learning, it has to be processed mindfully [13,15]. According to Kulhavy and Stock [41], when feedback is available to learners before they begin their memory search and construction of their answers, presearch availability exists. This causes mindless behavior by stopping the learner’s response–evaluation–adjustment process. On the contrary, confirming or disconfirming feedback (KR) should stimulate mindfulness.
Hence, learning settings that promote a student’s active generation of answers from his or her understanding and internal mental schemas, as it happens with MTF, should have a stronger influence than activities only showing information to the learner [40], as in KR or KCR.
However, the literature reports mixed results on multiple-try effects, so certain conditions are required for these principles to work. In addition, these theories do not consider the learner’s internal and emotional processes that could explain the obtained effects.

1.2. Motivation and Self-Determination

Emotions are vital for motivation, performance, and self-regulation [42,43]. Feedback influences both motivation and performance [18] as well as students’ regulation processes [36]. However, most research has focused on student performance, but few have covered feedback effects on motivation [4] and self-regulation [17].
The self-determination theory (SDT) [29,30] corresponds to a psychological model that can help describe the motivation construct and the emotional processes during learning. It has been effectively used in other learning situations, such as the use of gamification techniques [44,45], educational games [31,32], collaborative computer activities [46], and students’ conditions to achieve good grades [47].
SDT has its roots in the notion that curiosity and the desire to explore, discover, and comprehend the world are innate in human nature [29]. It understands the concept of intrinsic motivation as the natural inclination toward curiosity and interest in an activity. Three primary conditions or psychological needs facilitate its development: the individual feels competent, has autonomy, and feels related.
Competence refers to the result of exploration, learning, adaptation, and related processes in the interaction with the environment. In wider terms, competence can be understood as “the capacity for effective interactions with the environment that ensure the organism’s maintenance” [29]. For this reason, in 1959, White named it an effectance motivation [48]. He also states that the sensation resulting from competent actions is the reward per se. Hence, an innate satisfaction exists in training and broadening one’s abilities. This need for being effective in the environment’s interactions (competence) supplies the power for learning [48], and higher levels of self-efficacy have positive effects on student performance [49]. Diverse motivation theories propose that self-perceptions of competence and task processing are critical aspects of students’ motivation [4,50]. For that reason, situations generating feelings of higher competence will increase intrinsic motivation [29].
Autonomy involves experiencing free choice [33], where the locus of causality is perceived as internal rather than external. Although feelings of competence are essential for intrinsic motivation, they are not enough, and a sense of autonomy is required [29]. The third condition is relatedness, “the need to feel belongingness and connectedness with others” [33]. This dimension will not be considered in the present work, as the digital learning activity for the current research is individual, and the relatedness construct has proven to be important for internalization processes not covered in this study.
Cognitive evaluation theory (CET), the first of the six SDT sub-theories, focuses on the factors or conditions that stimulate or discourage intrinsic motivation. In this sense, when leveraging greater competence during an activity, diverse instructional elements such as feedback, rewards, or communications can increase intrinsic motivation toward that activity [51]. On the contrary, when the learner cannot be effective and attain a minimum level of competence, amotivation arises. CET states that learners need to perceive that their actions are self-determined and not imposed by external conditions [51]. For example, various studies have revealed that real rewards, pressured deadlines and evaluations, threatening communication and directives, and forced goals reduce intrinsic motivation because they all make the learner perceive the locus of causality as external [51]. From the SDT-CET perspective, learners’ intrinsic motivation is negatively affected whenever rewards are experienced as controlling, such as grades and other classroom rewards [29]. According to the authors, different characteristics of the learning environment (teacher’s attitude, existence of limits, grading system, etc.) can be perceived as controlling, negatively impacting the learner’s autonomy. However, they argue that it is possible to minimize these effects to the extent that they are presented in a more propositive and informative manner instead of being presented in a coercive, imposed, and controlling way.
Therefore, controlling conditions make learners feel pressure, which inhibits intrinsic motivation. For example, Ryan, Mims, and Koestner [52] found that controlling conditions positively correlated to expressed tension. Gottfried [53] reported that intrinsic motivation was negatively correlated to anxiety. The relationship between learners’ perceived competence and achievement is mediated by anxiety [27,54], and a negative correlation exists between anxiety and mathematics performance [55].
According to CET [30], an educational environment that fosters intrinsic motivation should provide a wide range of stimulating instructional elements, with adequate challenges, and an environment that promotes learner initiative and autonomy without controlling conditions and pressure. In this way, learning and achievement are more expected to arise.
Activities that generate optimal challenge and effort promote intrinsic motivation and self-determination [29]. When the student is in an activity that presents adequate levels of challenge, he or she will make a certain level of effort that will lead to feelings of competence. Therefore, optimal challenges will tend to generate higher levels of effort and competence. The level of effort can also be affected by controlling conditions [29]. A study by Ryan in 1982 [56] revealed that when subjects were controlled with evaluative feedback, they spent less effort, and their performance tended to be worse than those who received informational feedback. It seems that they accomplished what was requested but with less effort and in a poorer manner to react against the control.
When a learner has a high value for an activity, it shows that the regulation of that specific behavior (of performing such activity) has a high degree of internalization [30], meaning that it has been entirely accepted as one’s own, so one behaves, feels, and thinks that way and not because of rewards, punishments, or beliefs about society’s expected values [29].
Finally, there is scarce empirical research examining the SDT-CET dimensions considering feedback types and attempts (including MTF) under CBL settings. There is a need to advance the literature on how multiple-try inclusion affects students’ engagement, motivation, and related environmental conditions that foster motivation.

1.3. The Present Study

The present study aimed to explore the impacts of using MTF and KCR feedback under a CBL environment for primary students—three high-order math learning objectives were selected for exercising. The effects to explore were performance (learning improvements), motivation (in terms of interest, competence, and autonomy), and supportive learning conditions measured as effort, pressure (absence of), and value regarding the activity following the SDT-CET theory.

Research Hypothesis

Based on the literature review, positive effects should result from the use of multiple-try. In addition, these effects are expected to be greater than KCR for the selected learning objectives. Hence, we state the following:
Hypothesis 1.
Students learning in the educational game is higher in the Multiple-Try condition than in the KCR condition.
Better performance of MTF should be expected concerning promoting a greater perception of competence in students. Being a more active mechanism makes students relate their mental models with the wrong answer and adjust based on their knowledge. However, it is also possible that MTF generates frustration, trial-and-error behavior (mindlessness), and a lower feeling of competence if the learner’s errors increase and cannot be corrected. It will largely depend on the existence of an adequate level of challenge. By providing external information and being more passive, KCR could lower students’ sense of efficiency.
Regarding autonomy, higher levels could be expected in MTF than in KCR due to its interactive feature on the same question, which allows the solution or error-finding to be investigated more autonomously than if the correct answer was given immediately after the first response, as is the case of KCR. In this way, we anticipate the following:
Hypothesis 2.
Multiple-try generates in students higher motivation levels in terms of interest (H2a), competence (H2b), and autonomy (H2c) than KCR.
The concept of pressure is associated with conditions perceived as controlling by students that affect the degree of perceived autonomy. In this sense, KCR could be perceived as more controlling when indicating the correct answer after one attempt. And MTF could exert more pressure by not giving clear guidelines on continuing or correcting the question after an error. In this case, it would be pressure, not due to external control but rather feelings of loss of control or orientation. Thinking about the controller-informative dimension, MTF, to some degree, forces the person to try again after an error. That could be considered a controlling factor by some students.
Regarding the perceived effort, we believe that it will depend mainly on factors more related to an optimal level of challenge. In this sense, an adequate level of prior knowledge (medium-high) together with a high level of task complexity should generate an appropriate level of effort (medium-high) that causes a higher perception of competence. Then, in the particular case of MTF, its process of re-attempting the question without further information (KR) may require more effort when contrasted against KCR.
Regarding the perceived value, given that MTF generates a more active and profound reflection on errors and provides the opportunity to correct them, it is expected that this mechanism generates a greater sense of value than in the case of KCR. Such is especially true considering that the latter provides the correct answer externally to the individual, and as the activities are complex, concepts need to be understood rather than merely memorized. Based on the above, we consider the following:
Hypothesis 3.
Using multiple tries yields higher levels of students’ perception of effort (H3a), pressure (H3b), and value (H3c) than KCR.
Finally, there is no literature covering the effects of MTF and KCR from an emotional perspective, so it is not feasible to anticipate possible magnitudes for such differences if they occur.

2. Materials and Methods

2.1. Matematicast Learning Game

“MatematicaST” is a web platform with three mini-games to exercise diverse topics in mathematics. After logging in, students can select the game to be played each time. The question is displayed in the upper area and can be listened to when pressing the play button nearby (see Figure 1). The player has three lives per game, displayed as red hearts at the left. Each correct answer awards ten points, and the total points are on the upper-right. Also, on the right side is the list with the highest scores.
The “Numbers Identification” mini-game determines a number’s symbolic representation according to its place value (Learning Objective 1). The information is provided in units, tens, and hundreds, while four alternatives are presented as boxes. Options are built up by combining the digits shown in the question.
The “Numbers Ordering” mini-game asks to order four numbers from smallest to largest within 10 to 1000 (Learning Objective 2). Numbers appear in disorder at the top inside a box (See Figure 2a). Boxes can be dragged to an empty place and confirm the answer by pressing the “ESTE ES EL ORDEN” button (this is the order). This button will only evaluate the order of the boxes once all the empty places have a number.
The “Money Counting” mini-game (Figure 2b) requires counting coins of a different amount (Learning Objective 3). Coins are limited to one instance maximum for 5, 50, and 500 coins, while up to four instances for 1, 10, and 100. Figure 2b shows an exercise displaying one 50-coin, three 10-coins, and four 1-coins (coins are in Chilean peso—CLP).

MTF and KCR Game Versions

The KCR version for all three mini-games provides immediate feedback to the student after each answer. When answering correctly, the game presented to the user a motivating message chosen randomly between samples as “Buena respuesta!” (Good answer!) or “Sigue así!” (keep it up!), which remained on the screen for 1.1 s before going to the next exercise (See Figure 3a). If the selected answer was incorrect, the game presented a message indicating so, including the correct answer and the text “keep trying” (Figure 3b). The game displays a continue button to the next exercise after 2.2 s of showing the message. In addition, the game provided sounds to complement the correct/incorrect feedback.
The MTF version gives the student three attempts per exercise, with KR among attempts. They are displayed graphically below the life section by using three yellow dices. The student obtains total points and has no life discounted when answering correctly within those three attempts. The game reduces one life when losing and goes to the next exercise without providing the correct answer. It also displays the selected incorrect alternatives in red boxes as a memory aid for previously chosen options. Figure 4 shows the Money counting mini-game with one attempt left after answering incorrectly 384 and 438 while having one 500-, three 100-, four 10-, and three 1-coins. No additional hints, clues, or strategies were provided.

2.2. Participants and Design

Students from four courses (3° and 4° grade) from two schools in the region of Valparaíso, Chile, participated in the intervention. Initially, we anticipated the participation of 84 students. However, some students were absent on the activity days, did not have their parents’ signed consent, or did not complete all the instruments (post-test or post-IMI), so the final number of participants was 73.
The study considered a randomized pre-test–post-test design, with feedback type as a factor with two conditions: a multiple-try with KR and a single-try with KCR. Students were randomly assigned to a condition with 36 (13 male and 23 female) for Multiple-try and 37 (14 male and 23 female) for the KCR condition.

2.3. Instruments

2.3.1. Math Tests

Pre- and post-tests were built up following the three learning objectives considered from the Chilean mathematics program and its guide textbook. Teachers from the two schools were involved in test construction and validated their instructions, exercises, and conditions. Pre-test and post-test considered six exercises for each learning objective (number identification, number order, and money counting), with a total of 9 points.

2.3.2. IMI Questionnaire

The Intrinsic Motivation Inventory (IMI) [57] is a measurement instrument for evaluating the SDT-CET constructs in subjects participating in a given activity. The tool has been applied in various studies concerning intrinsic motivation [29,30]. The present research considered six constructs: interest, perceived competence, pressure, effort, perceived choice (autonomy), and perceived value. Perception tests had 16 sentences with a 5-point Likert scale (from 1 “strongly disagree” to 5 “strongly agree”). All dimensions had three sentences each except for pressure and effort, with only two items each. We translated from English to Spanish and validated it with the schoolteachers, adjusting some vocabulary. For example, sentences using the words tense or pressure were changed to the word nervous, as students did not understand those first words.

2.4. Procedure

In the beginning, students received a motivational talk with the general purpose of the experiment. Some instructions were also explained regarding the procedure and instruments (tests) to be applied (10 min). Then, students completed the pre-IMI test and the learning pre-test, both in a paper-and-pencil format for 15 min. After that, they logged into the web platform and played the games in the school computer laboratory for 55 min. After playing, they answered the post-IMI test to measure their self-report of motivation, pressure, effort, and utility. Students completed the learning post-test not only after finishing the game activity but also after a 15-minute recess period to allow them to rest.
The learning objectives involved in the study were already taught and evaluated several weeks before the experiment. Therefore, the games’ playing corresponded to an exercise of previous knowledge, voluntary and without any external motivators regarding school points or grades.

3. Results

Data was processed with SPSS software version 29 for the statistical analyses and graphs generation. Partial eta squared (η2) was used as a measure of effect size.

3.1. MTF and KCR Learnings

First, Table 1 shows the descriptive statistics for pre- and post-test scores per condition. A 2-way repeated measures ANOVA was conducted to evaluate the effects of test type (pre-test/post-test) and condition (MTF/KCR). The simple main effect of test type was significant with F(1, 71) = 27.064, p < 0.001, η2 = 0.276, meaning that overall, post-tests scored higher than pre-tests. The two-way interaction between test type and condition was also significant with F(1, 71) = 8.096, p = 0.006, η2 = 0.102. Now, when analyzing the effect of condition on each test type, significant differences were obtained for the post-test F(1, 71) = 4.800, p = 0.032, η2 = 0.06 but not for the pre-test F(1, 71) = 0.254, p = 0.62, η2 = 0.004, meaning that pre-test scores were not significantly different among MTF and KCR conditions, but their post-test scores were significantly different, as Figure 5a shows.
Then, the effect of each test type was analyzed at each condition, showing that significant differences existed for MTF (F(1, 71) = 31.944, p < 0.001, η2 = 0.31) but not for KCR (F(1, 71) = 2.816, p = 0.098, η2 = 0.038), meaning that significant learning occurred in the MTF condition but not in the KCR one.
In addition, a 1-way ANCOVA was carried out to determine the effect of condition (MTF vs. KCR) on learning gains (post-test–pre-test scores) after controlling for the pre-test score of students, showing a significant difference between the conditions with F(1, 70) = 12.481, p < 0.001, η2 = 0.15. This means that MTF outperforms KCR in learning gains after adjustment for pre-test, as Figure 5b depicts.

3.2. MTF and Motivation

Table 2 shows the descriptive statistics of IMI dimensions per condition. Pre- and post-IMI test values are considered measures for the dimensions of interest, perceived competence, perceived choice, effort, no pressure (reverse of pressure), and value. A Cronbach alpha of 0.74 for pre-IMI and 0.68 for post-IMI perception tests were obtained.
First, the post-IMI values were analyzed. A 2-way repeated measures ANOVA was carried out to evaluate the effects of conditions (MTF/KCR) and IMI dimensions on test values, resulting in a significant interaction (F(5, 67) = 2.422, p = 0.044, η2 = 0.153). Pairwise comparisons under Bonferroni adjustment for multiple comparisons contrasted both experimental conditions for each IMI construct. Significant differences for No-Pressure (F(1, 71) = 6.210, p = 0.015, η2 = 0.080) and Choice (F(1, 71) = 3.442, p = 0.068, η2 = 0.046) were obtained. This means that MTF presented higher levels of pressure and perceived choice than KCR on post-test values (see Figure 6).
Secondly, were analyzed the differences between post-IMI and pre-IMI values. A 2-way repeated measures ANOVA was carried out to evaluate the effects of conditions (MTF/KCR) and the six IMI dimensions on the pre-post IMI delta values. Results show a significant interaction with F(5, 67) = 3.394, p = 0.009, η2 = 0.202). Pairwise comparisons under Bonferroni correction contrasted both experimental conditions for each IMI-CET construct. Now, significant differences were found for the dimensions of Competence (F(1, 71) = 4.445, p = 0.039, η2 = 0.059), No-Pressure (F(1, 71) = 5.673, p = 0.020, η2 = 0.074), and Choice (F(1, 71) = 4.363, p = 0.040, η2 = 0.058). Therefore, MTF presented significantly higher increases in perceived competence and perceived autonomy (choice) than KCR but also presented a significantly lower gain on the No-Pressure dimension (see Figure 7).

4. Discussion

4.1. H1: Students Learning in the Educational Game Is Higher in the Multiple-Try Condition Than in the KCR Condition

First, results show that both conditions provided learning gains; however, only in the MTF group were they significant. It is especially relevant considering that the intervention involved a lapse of 55 min of playing with the games only. According to what was expected in H1, and despite the limited intervention time, MTF rose with higher learning improvements. In addition, such differences were mostly due to post-test differences as pre-test scores per condition did not differ significantly.
Our MTF implementation reported similar effects as Van der Kleij et al. [13] meta-analysis for elaborated feedback, higher than Clariana and Koul [16] reporting an ES 0.11 for MTF under high-order LOs, but lower improvements than Attali’s [8] MTC and MTH implementations.
Initially, one might think that KCR feedback should have given better results in learning gains than MTF. The first provides more direct and immediate information about the correct answer than the other alternative. On the contrary, results suggest that the MTF exploratory process allowed a more hands-on role for the student to think of the correct answer, as MTF did not directly provide it but only knew that the chosen alternative was incorrect (KR), thus allowing for a more active approach that led to deeper learning. This result is consistent with the idea that multiple-try “provides an opportunity for elaboration and reorganization of information that may be beneficial for learning” [8].
However, for this information reorganization to occur, you need learners to have a minimum level of prior knowledge (PK) to work internally. In this sense, the students in our experiment presented average pre-test levels of around 80%. These prior knowledge ranges are consistent with the literature reviewed that had good results using MTF [8,58]. In any case, additional studies are needed to more precisely establish the PK levels required for efficient use of the different types of feedback and, in particular, of multiple-try.
A relevant issue is the MTF implementation. To the best of our knowledge, this study is the first considering under a CBL context an MTF with three attempts + KR and no KCR nor EF at the end. No hints nor correct responses were provided even after the third try. This fact, together with the appropriate prior knowledge level, could have discouraged students from following a trial-and-error strategy [58] or an automatic and mindless behavior [17] as presearch availability [41] was avoided.
Furthermore, the idea of not presenting the correct answer after a first attempt is not new. In the context of intelligent tutoring systems (ITS), the informative tutoring feedback model (ITF) [4,59] involves providing different components without explicitly indicating the correct answer. It seeks to guide students to identify and correct errors by applying more effective learning strategies.
Therefore, more studies on the effects of using multiple tries with a diverse number of attempts, prior knowledge levels, and interaction mechanics are required. Also, to consider how the effects of multiple-try are affected when using item types different from multiple choice.

4.2. H2: Multiple-Try Generates in Students Higher Motivation Levels in Terms of Interest (H2a), Competence (H2b), and Autonomy (H2c) Than KCR

When analyzing H2 and students’ perception of motivation, our study revealed very high levels of interest and competence during the activity. Also, positive delta values on pre- and post-test IMI measures for both conditions were obtained on interest, perceived competence, and choice (autonomy), meaning that the activity enhanced those feelings. In addition, the differences (increases) in competence and choice were significantly higher for MTF when contrasted against KCR. Therefore, MTF outperformed KCR on two out of three of the most relevant dimensions supporting motivation, which are competence (H2b) and autonomy (H2c), but not in the dimension of interest (H2a).
Results are consistent with previous research demonstrating that motivational variables were stronger predictors of achievement in mathematics [60]. The implemented gamification mechanisms (points, lives, and leaderboard) providing progression and competition dynamics seemed to have helped maintain high engagement and motivation levels, minimizing emotional differences across conditions.
The implemented MTF only provided KR (and not KCR at the end as in other studies); it seemed to have a cost in interest when contrasted to providing explicit content feedback, such as in KCR, in terms of not presenting a significant difference. Despite the above, MTF presented higher competence perception effects than KCR. Firstly, the activity seems to have generated an optimal sense of challenge [29] that helped the development of larger feelings of competence for both conditions (MTF and KCR).
Multiple-try also presented higher effects than KCR on perceptions of autonomy. It seems to confirm that the MTF characteristic of iterating over the same question when answering wrongly without giving the correct answer activates internal processes that are beneficial for learning as it does not interrupt the response–evaluation–adjustment process with external information [41].
From a certain point of view, MTF could act as relatively negative feedback, thus affecting competence since it is activated upon error and, in our case, provides only the response correctness (KR). However, the degree of involvement in the task, sense of challenge, and commitment were high enough to overcome the negative feelings and transform the wrong answer into a motivating situation since the additional attempts of the MTF allowed the student to discover on their own (sense of autonomy) the mistakes and correct them immediately and for a next time (sense of competence).
External rewards negatively affect the sense of autonomy. Within our experiment, no real incentives were given, such as marks or points that could be used academically. Therefore, a fictional environment was created in which to explore and learn without direct consequences in the real world, which promoted freedom and autonomy during the game.
An interesting result was that despite presenting higher levels of competence and autonomy, MTF did not generate higher levels of interest than KCR. A possible explanation could be that the intervention duration may have affected or, more likely, the fact that the interactive mechanism provided by MTF + KR did not make the game more interesting than a single attempt + KCR. The instructional material and the rest of the game were the same: history, context, exercises, and gamification elements such as points and ranking. So, the control condition selection was very relevant because presumably, if MTF were compared against traditional classes or an online pdf with exercises, the results would have been different and in favor of the game without being possible to distinguish the effects of the MTF from those of the game itself.

4.3. H3: Using Multiple Tries Yields Higher Levels of Students’ Perception of Effort (H3a), Pressure (H3b), and Value (H3c) Than KCR

As mentioned above, there must be a suitable environment and students’ attitudes toward learning. It was evaluated using the dimensions of effort, pressure, and value. H3 anticipated higher levels of students’ perception of effort (H3a), pressure (H3b), and value (H3c) for the MTF condition. In general, results revealed high levels for the value dimension while medium levels for effort and no pressure across conditions. MTF presented a significantly higher post-test score than KCR in the no-pressure dimension. Regarding pre- and post-tests, IMI differences only effort presented a decrease (negative delta) for both conditions, meaning that students expected to make more effort than they needed. KCR presented a significantly higher difference in the no-pressure construct but no significant differences in effort and value when contrasted against MTF. Therefore, H3 was partially accepted, confirming H3b but rejecting H3a and H3c. We expected higher levels of effort and pressure for students under MTF because they had a long and indirect way of arriving at the correct answers (via the attempts in case of a previous failure) when contrasted to KCR. Also, our implemented MTF did not provide much guidance or information, supposing more difficulty and pressure to find the correct answer. In this sense, it would be interesting to study what would happen to the levels of effort and pressure when combining MTF with KCR or hints such as those implemented in Atali [8]. The hints consist of partial pieces of information that helped solve the math problem but without providing the result explicitly, as happens in KCR.
External regulations generate less interest, value, and effort toward the activity, delegating the success responsibility to others [51]. During the present experience, the levels of effort, no-pressure, and value constructs were adequately high for both conditions, giving general indications of learners’ high commitment, self-determination, and mindfulness.
Pressure is also a measure of external coercion or controlling conditions (and lack of autonomy), generating less intrinsic motivation. Therefore, the high levels of no-pressure (H3b) throughout the experiment support the idea of little existence of coercive or controlling elements, thus promoting intrinsic motivation.
The effort levels (H3a) were relatively medium-high throughout the entire experience, reinforcing the impression that there was an adequate level of challenge: the exercises’ difficulty level versus students’ ability level, especially in terms of prior knowledge. Likewise, the high level of effort reported indicates the small existence of factors perceived as controlling. There was no evidence of less effort as a reaction to external control by students in contrast to what was reported by Ryan in his study [56].
Another factor that could affect the level of effort is the mechanism for generating item alternatives. The existence of easily discarded or obvious options was avoided. For this, part of the digits of the correct answer or permutations of said digits were used to construct the remaining alternatives to deliver options close to the correct one but which were not. For example, when faced with the number representing seven units and three hundred, the alternatives were 307, 37, 370, and 703. In this way, students’ sense of challenge and effort was reinforced, but without taking them out of their comfort zone, which would have generated frustration and pressure. Therefore, these challenging alternatives promoted more active attention and a greater sense of competence after solving them correctly.
In our MTF, responses to earlier attempts for the same exercise were provided to students as a visual aid to help them remember previous answers. It could have reduced the effort required to arrive at the correct answer via the attempts while diminishing the frustration or pressure when going trial after trial. We think that the exercises’ complexity level and item type could have moderated such differences, as choosing from four alternatives is different from a constructed response (or fill-in-the-blank question) in a similar way as reported by Attali and Van der Kleij [61].
Regarding the value dimension (H3c), the intervention presented high levels before and after the experience, indicating that the students found it helpful and implying high levels of self-determination and self-regulation for the whole group. On value per condition, no significant differences were observed. Thus, H3c was rejected. Students perceived the multiple-try as valuable as KCR despite significant differences in learning. To a certain extent, MTF students did not acknowledge their progress and improvement the same way as KCR students did. Maybe more time was needed to internalize the improvements and lessons learned. Then, it would be interesting to see the effects of incorporating delayed summary feedback to improve self-reflection processes.

5. Conclusions

The present study explored the impacts of using multiple-try feedback (MTF) and Knowledge of Correct Response Feedback (KCR) in a computer-based learning (CBL) environment for third and fourth-primary students, focusing on 3-digit number identification, ordering, and money counting as part of the math national curriculum. The research aimed to assess performance, motivation, and supportive learning conditions based on self-determination theory (SDT) and its cognitive evaluation sub-theory (CET).
Results indicate that MTF led to significant learning gains, outperforming KCR in the high-order math learning objectives involved. The iterative process of MTF allowed an active approach that positively influenced students’ confidence in their abilities, leading to deeper learning. MTF encouraged student engagement, allowing them to explore, learn from mistakes, and adjust their understanding independently, fostering a sense of autonomy. Also, MTF presented higher levels of pressure due to the iterative process and the absence of immediate correct answers. However, the pressure was not excessive and seemed to not negatively impact intrinsic motivation.
The experimental setup effectively managed task complexity and students’ prior knowledge. This control facilitated positive outcomes related to the use of multiple-try feedback. The study supports previous research indicating that multiple-try feedback is beneficial for high-order learning outcomes in text-comprehension contexts. Moreover, this study extends these findings to the K-12 math domain.
The MatematicaST educational platform proved to be an effective medium for math learning. The games maintained high levels of student interest and competence during the activity, together with relatively high effort levels, indicating a suitable level of challenge during the learning activities. Overall, the training activity was perceived as valuable by students, indicating that they found the learning activities meaningful and helpful. All of the above resulted in high learning gains in both conditions.
The study utilized various instruments, including pre- and post-tests for learning objectives and the Intrinsic Motivation Inventory (IMI) questionnaire. The inclusion of both pre- and post-IMI questionnaires allowed for a comprehensive analysis of students’ self-perceptions regarding the activity. This approach provided valuable insights into students’ perspectives, which would not have been achievable with the post-IMI questionnaire alone.
The study involved primary students from third and fourth grade from two schools in the region of Valparaíso. They showed high motivation during the activity, which resulted in active and meaningful participation. However, the short nature of the training could have limited the time for students to internalize their progress and learning. Therefore, future research should consider longer interventions and assess diverse outcomes, especially regarding students’ perceived usefulness.
Other future research directions should include exploring the combination of MTF with other feedback types, such as KCR, hints, or delayed summary feedback, as they could enhance students’ self-reflection processes and further improve learning outcomes. Also, to research deeper on the levels of prior knowledge and learning objective complexity suitable for effective utilization of multiple-try feedback. There is a lack of gender-specific studies examining differences in the use of multiple-try feedback and its outcomes. Further research in this area is necessary.

Author Contributions

Conceptualization, C.C. and S.R.; methodology, C.C. and S.R.; software, C.C.; validation, C.C., S.R. and D.C.-P.; formal analysis, C.C.; investigation, C.C.; resources, C.C.; data curation, C.C.; writing—original draft preparation, C.C.; writing—review and editing, C.C. and D.C.-P.; visualization, S.R.; supervision, S.R.; project administration, C.C.; funding acquisition, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Pontifical Catholic University of Valparaíso grant number DI039.416/2021.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Bioethics and Biosecurity Committee of Pontificia Universidad Católica de Valparaíso (protocol code BIOEPUCV-H 659-2023 on 29 June 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available as it may present identification data of the study subjects.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tsai, Y.-L.; Tsai, C.-C. A meta-analysis of research on digital game-based science learning. J. Comput. Assist. Learn. 2020, 36, 280–294. [Google Scholar] [CrossRef]
  2. Hussein, M.H.; Ow, S.H.; Cheong, L.S.; Thong, M.; Ebrahim, N.A. Effects of Digital Game-Based Learning on Elementary Science Learning: A Systematic Review. IEEE Access 2019, 7, 62465–62478. [Google Scholar] [CrossRef]
  3. Cheung, A.; Slavin, R. The effectiveness of educational technology applications for enhancing mathematics achievement in K-12 classrooms: A meta-analysis. Educ. Res. Rev. 2013, 9, 88–113. [Google Scholar] [CrossRef]
  4. Narciss, S. Feedback strategies for interactive learning tasks. In Handbook of Research on Educational Communications and Technology, 3rd ed.; Spector, J.M., Merril, M.D., van Merriënboer, J.J.G., Driscoll, M.P., Eds.; Lawrence Erlbaum: Mahwah, NJ, USA, 2008; pp. 125–144. [Google Scholar]
  5. Pressey, S.L. A simple apparatus which gives tests and scores e and teaches. Sch. Soc. 1926, 23, 373–376. [Google Scholar]
  6. Pressey, S.L. Development and Appraisal of Devices Providing Immediate Automatic Scoring of Objective Tests and Concomitant Self-Instruction. J. Psychol. 1950, 29, 417–447. [Google Scholar] [CrossRef]
  7. Faulconer, E.; Griffith, J.C.; Frank, H. If at first you do not succeed: Student behavior when provided feedforward with multiple trials for online summative assessments. In Teaching in Higher Education; Taylor & Francis: Abingdon, UK, 2019. [Google Scholar] [CrossRef]
  8. Attali, Y. Effects of multiple-try feedback and question type during mathematics problem solving on performance in similar problems. Comput. Educ. 2015, 86, 260–267. [Google Scholar] [CrossRef]
  9. Azevedo, R.; Bernard, R.M. A meta-analysis of the effects of feedback in computer-based instruction. J. Educ. Comput. Res. 1995, 13, 111–127. [Google Scholar] [CrossRef]
  10. Black, P.; Wiliam, D. Assessment and classroom learning. Assess. Educ. Princ. Policy Pract. 1998, 5, 7–74. [Google Scholar] [CrossRef]
  11. Bloom, B.S. The 2 sigma problem: The search for methods of group instruction as effective as one-to-one tutoring. Educ. Res. 1984, 13, 4–16. [Google Scholar] [CrossRef]
  12. Jaehnig, W.; Miller, M. Feedback types in programmed instruction: A systematic review. Psychol. Rec. 2007, 57, 219–232. [Google Scholar] [CrossRef]
  13. Van der Kleij, F.M.; Feskens, R.C.W.; Eggen, T.J.H.M. Effects of Feedback in a Computer-Based Learning Environment on Students’ Learning Outcomes: A Meta-Analysis. Rev. Educ. Res. 2015, 85, 475–511. [Google Scholar] [CrossRef]
  14. Shute, V.J. Focus on Formative Feedback. Rev. Educ. Res. 2008, 78, 153–189. [Google Scholar] [CrossRef]
  15. Bangert-Drowns, R.L.; Kulik, C.-L.C.; Kulik, J.A.; Morgan, M. The Instructional Effect of Feedback in Test-Like Events. Rev. Educ. Res. 1991, 61, 213–238. [Google Scholar] [CrossRef]
  16. Clariana, R.B.; Koul, R. Multiple-Try Feedback and Higher-Order Learning Outcomes. Int. J. Instr. Media 2005, 32, 239. [Google Scholar]
  17. Van der Kleij, F.M.; Timmers, C.F.; Eggen, T.J.H.M. The effectiveness of methods for providing written feedback through a computer-based assessment for learning: A systematic review. CADMO 2011, 19, 21–39. [Google Scholar] [CrossRef]
  18. Kluger, A.; DeNisi, A.S. The effects of feedback interventions on performance: A historical review, a meta-analysis, and a preliminary feedback intervention theory. Psychol. Bull. 1996, 119, 254–284. [Google Scholar] [CrossRef]
  19. Candel, C.; Vidal-Abarca, E.; Cerdán, R.; Lippmann, M.; Narciss, S. Effects of timing of formative feedback in computer-assisted learning environments. J. Comput. Assist. Learn. 2020, 36, 718–728. [Google Scholar] [CrossRef]
  20. Agina, A.; Kommers, P.; Heylen, D. Towards Understanding Human-Media Interaction: The Effect of the Computer’s Answer-Until-Correct (AUC) vs. Knowledge-of-Result (KR) Task Feedback on Young Users’ Behavioral Interaction Development through a Digital Playground. J. Child. Dev. Disord. 2016, 2, 100011. [Google Scholar] [CrossRef]
  21. Stevenson, C.E.; Hickendorff, M. Learning to solve figural matrix analogies: The paths children take. Learn. Individ. Differ. 2018, 66, 16–28. [Google Scholar] [CrossRef]
  22. Fyfe, E.R. Providing feedback on computer-based algebra homework in middle-school classrooms. Comput. Hum. Behav. 2016, 63, 568–574. [Google Scholar] [CrossRef]
  23. Murphy, P. Reading comprehension exercises online: The effects of feedback, proficiency and interaction. Lang. Learn. Technol. 2007, 3, 107–129. [Google Scholar]
  24. Fyfe, E.R.; Rittle-Johnson, B.; DeCaro, M.S. The effects of feedback during exploratory mathematics problem solving: Prior knowledge matters. J. Educ. Psychol. 2012, 104, 1094–1108. [Google Scholar] [CrossRef]
  25. Clariana, R.B.; Lee, D. The effects of recognition and recall study tasks with feedback in a computer-based vocabulary lesson. Educ. Technol. Res. Dev. 2001, 49, 23–36. [Google Scholar] [CrossRef]
  26. Peixoto, F.; Sanches, C.; Mata, L.; Monteiro, V. “How do you feel about math?”: Relationships between competence and value appraisals, achievement emotions and academic achievement. Eur. J. Psychol. Educ. 2017, 32, 385–405. [Google Scholar] [CrossRef]
  27. Pérez-Fuentes, M.D.C.; Núñez, A.; Molero, M.D.M.; Gázquez, J.J.; Rosário, P.; Núñez, J.C. The Role of Anxiety in the Relationship between Self-efficacy and Math Achievement. Psicol. Educ. 2020, 26, 137–143. [Google Scholar] [CrossRef]
  28. Chen, C.; Law, V.; Huang, K. The roles of engagement and competition on learner’s performance and motivation in game-based science learning. Educ. Technol. Res. Dev. 2019, 67, 1003–1024. [Google Scholar] [CrossRef]
  29. Deci, E.L.; Ryan, R.M. Intrinsic Motivation and Self-Determination in Human Behavior; Plenum Press: New York, NY, USA, 1985. [Google Scholar]
  30. Ryan, R.M.; Deci, E.L. Intrinsic and extrinsic motivation from a self-determination theory perspective: Definitions, theory, practices, and future directions. Contemp. Educ. Psychol. 2020, 61, 101860. [Google Scholar] [CrossRef]
  31. Liu, Y.C.; Wang, W.-T.; Lee, T.-L. An Integrated View of Information Feedback, Game Quality, and Autonomous Motivation for Evaluating Game-Based Learning Effectiveness. J. Educ. Comput. Res. 2020, 59, 3–40. [Google Scholar] [CrossRef]
  32. Chen, C.-H.; Law, V. Scaffolding individual and collaborative game-based learning in learning performance and intrinsic motivation. Comput. Hum. Behav. 2016, 55B, 1201–1212. [Google Scholar] [CrossRef]
  33. Deci, E.; Ryan, R. The “What” and “Why” of Goal Pursuits: Human Needs and the Self-Determination of Behavior. Psychol. Inq. 2000, 11, 227–268. [Google Scholar] [CrossRef]
  34. Andre, T.; Thieman, A. Level of adjunct question, type of feedback, and learning concepts by reading. Contemp. Educ. Psychol. 1988, 13, 296–307. [Google Scholar] [CrossRef]
  35. Barana, A.; Marchisio, M.; Sacchet, M. Interactive Feedback for Learning Mathematics in a Digital Learning Environment. Educ. Sci. 2021, 11, 279. [Google Scholar] [CrossRef]
  36. Hattie, J.; Timperley, H. The power of feedback. Rev. Educ. Res. 2007, 77, 81–112. [Google Scholar] [CrossRef]
  37. Dempsey, J.V. Interactive Instruction and Feedback; Educational Technology: Washington, DC, USA, 1993. [Google Scholar]
  38. Timmers, C.; Veldkamp, B. Attention paid to feedback provided by a computer-based assessment for learning on information literacy. Comput. Educ. 2011, 56, 923–930. [Google Scholar] [CrossRef]
  39. Tsai, F.-H.; Tsai, C.-C.; Lin, K.-Y. The evaluation of different gaming modes and feedback types on game-based formative assessment in an online learning environment. Comput. Educ. 2015, 81, 259–269. [Google Scholar] [CrossRef]
  40. Clariana, R.B.; Wagner, D.; Murphy, L.C.R. Applying a Connectionist Description of Feedback Timing. Educ. Technol. Res. Dev. 2000, 48, 5–21. [Google Scholar] [CrossRef]
  41. Kulhavy, R.W.; Stock, W.A. Feedback in written instruction: The place of response certitude. Educ. Psychol. Rev. 1989, 1, 279–308. [Google Scholar] [CrossRef]
  42. Arguedas, M.; Daradoumis, T. Analysing the role of a pedagogical agent in psychological and cognitive preparatory activities. J. Comput. Assist. Learn. 2021, 37, 1167–1180. [Google Scholar] [CrossRef]
  43. Dunn, T.J.; Kennedy, M. Technology Enhanced Learning in higher education; motivations, engagement and academic achievement. Comput. Educ. 2019, 137, 104–113. [Google Scholar] [CrossRef]
  44. Proulx, J.-N.; Romero, M.; Arnab, S. Learning Mechanics and Game Mechanics Under the Perspective of Self-Determination Theory to Foster Motivation in Digital Game Based Learning. Simul. Gaming 2017, 48, 81–97. [Google Scholar] [CrossRef]
  45. Sailer, M.; Hense, J.U.; Mayr, S.K.; Mandl, H. How gamification motivates: An experimental study of the effects of specific game design elements on psychological need satisfaction. Comput. Hum. Behav. 2017, 69, 371–380. [Google Scholar] [CrossRef]
  46. Liao, C.W.; Chen, C.H.; Shih, S.J. The interactivity of video and collaboration for learning achievement, intrinsic motivation, cognitive load, and behavior patterns in a digital game-based learning environment. Comput. Educ. 2019, 133, 43–55. [Google Scholar] [CrossRef]
  47. Martinek, D.; Zumbach, J.; Carmignola, M. How Much Pressure Do Students Need to Achieve Good Grades?—The Relevance of Autonomy Support and School-Related Pressure for Vitality, Contentment with, and Performance in School. Educ. Sci. 2022, 12, 510. [Google Scholar] [CrossRef]
  48. White, R.W. Motivation reconsidered: The concept of competence. Psychol. Rev. 1958, 66, 297–333. [Google Scholar] [CrossRef] [PubMed]
  49. Rodriguez, M. The Role of Classroom Assessment in Student Performance on TIMSS. In Applied Measurement in Education; Taylor & Francis: Abingdon, UK, 2004; Volume 17, pp. 1–24. [Google Scholar] [CrossRef]
  50. Pintrich, P.R. Motivation and classroom learning. In Handbook of Psychology. Vol. 7. Educational Psychology; Reynolds, W.M., Miller, G.E., Eds.; John Wiley & Sons: Hoboken, NJ, USA, 2003; pp. 103–122. [Google Scholar]
  51. Ryan, R.; Deci, E. Self-determination theory and the facilitation of intrinsic motivation, social development, and well-being. Am. Psychol. 2000, 55, 68–78. [Google Scholar] [CrossRef] [PubMed]
  52. Ryan, R.M.; Mims, V.; Koestner, R. Relation of reward contingency and interpersonal context to intrinsic motivation: A review and test using cognitive evaluation theory. J. Pers. Soc. Psychol. 1983, 45, 736–750. [Google Scholar] [CrossRef]
  53. Gottfried, A.E. Relationships between academic intrinsic motivation and anxiety in children and young adolescents. J. Sch. Psychol. 1982, 20, 205–215. [Google Scholar] [CrossRef]
  54. Wang, L. Mediation Relationships Among Gender, Spatial Ability, Math Anxiety, and Math Achievement. Educ. Psychol. Rev. 2020, 32, 362–370. [Google Scholar] [CrossRef]
  55. Namkung, J.M.; Peng, P.; Lin, X. The relation between mathematics anxiety and mathematics performance among school-aged students: A meta-analysis. Rev. Educ. Res. 2019, 89, 459–496. [Google Scholar] [CrossRef]
  56. Ryan, R.M. Control and information in the intrapersonal sphere: An extension of cognitive evaluation theory. J. Pers. Soc. Psychol. 1982, 43, 450–461. [Google Scholar] [CrossRef]
  57. Monteiro, V.; Mata, L.; Peixoto, F. Intrinsic Motivation Inventory: Psychometric Properties in the Context of First Language and Mathematics Learning. Psicol. Reflex. Crit. 2015, 28, 434–443. [Google Scholar] [CrossRef]
  58. Attali, Y.; Laitusis, C.; Stone, E. Differences in Reaction to Immediate Feedback and Opportunity to Revise Answers for Multiple-Choice and Open-Ended Questions. Educ. Psychol. Meas. 2016, 76, 787–802. [Google Scholar] [CrossRef] [PubMed]
  59. Merrill, D.C.; Reiser, B.J.; Merrill, S.K.; Landes, S. Tutoring: Guided learning by doing. Cognit. Instruct. 1995, 13, 315–372. [Google Scholar] [CrossRef]
  60. Abín, A.; Núñez, J.C.; Rodríguez, C.; Cueli, M.; García, T.; Rosário, P. Predicting Mathematics Achievement in Secondary Education: The Role of Cognitive, Motivational, and Emotional Variables. Front. Psychol. 2020, 11, 876. [Google Scholar] [CrossRef] [PubMed]
  61. Attali, Y.; Van der Kleij, F. Effects of feedback elaboration and feedback timing during computer-based practice in mathematics problem solving. Comput. Educ. 2017, 110, 154–169. [Google Scholar] [CrossRef]
Figure 1. Number identification mini-game.
Figure 1. Number identification mini-game.
Education 13 01119 g001
Figure 2. (a) Number ordering mini-game; (b) money counting mini-game. Both are in their base versions.
Figure 2. (a) Number ordering mini-game; (b) money counting mini-game. Both are in their base versions.
Education 13 01119 g002
Figure 3. Numbers ordering mini-game. (a) Feedback for the correct answer; (b) feedback upon a wrong answer.
Figure 3. Numbers ordering mini-game. (a) Feedback for the correct answer; (b) feedback upon a wrong answer.
Education 13 01119 g003
Figure 4. The multiple-try version of the money counting mini-game.
Figure 4. The multiple-try version of the money counting mini-game.
Education 13 01119 g004
Figure 5. (a) Pre- and post-test scores per condition; (b) learning gains per condition after controlling for pre-test.
Figure 5. (a) Pre- and post-test scores per condition; (b) learning gains per condition after controlling for pre-test.
Education 13 01119 g005
Figure 6. Post-perception scores of IMI constructs per experimental condition.
Figure 6. Post-perception scores of IMI constructs per experimental condition.
Education 13 01119 g006
Figure 7. Pre- and post-perception differences of IMI constructs per experimental condition.
Figure 7. Pre- and post-perception differences of IMI constructs per experimental condition.
Education 13 01119 g007
Table 1. Pre- and post-test descriptive statistics per condition.
Table 1. Pre- and post-test descriptive statistics per condition.
ConditionMeanStd. DeviationN
Pre-testMTF7.172.2136
KCR7.411.8337
Total7.292.0173
Post-testMTF8.530.9336
KCR7.801.7637
Total8.161.4573
Table 2. IMI perception constructs’ descriptive statistics per condition.
Table 2. IMI perception constructs’ descriptive statistics per condition.
ConstructConditionPre-TestPost-TestDeltaN
MeanSDMeanSDMeanSD
InterestMTF4.570.584.860.350.290.6636
KCR4.580.604.780.440.200.6537
Total4.580.594.820.400.240.6673
CompetenceMTF4.000.804.620.560.600.6736
KCR4.200.764.460.760.270.6937
Total4.100.784.540.670.430.7073
EffortMTF4.570.804.240.87−0.331.1036
KCR4.430.864.070.91−0.361.1537
Total4.500.834.150.88−0.351.1273
No-PressureMTF3.880.943.891.060.011.1936
KCR3.880.904.430.780.550.7037
Total3.880.914.160.960.291.0073
ChoiceMTF3.401.114.210.740.801.0836
KCR3.581.073.880.800.291.0237
Total3.491.094.040.790.541.0773
ValueMTF4.550.584.760.420.220.5636
KCR4.390.854.620.750.230.7437
Total4.470.734.690.610.220.6573
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cubillos, C.; Roncagliolo, S.; Cabrera-Paniagua, D. Learning and Motivation When Using Multiple-Try in a Digital Game for Primary Students in Chile. Educ. Sci. 2023, 13, 1119. https://doi.org/10.3390/educsci13111119

AMA Style

Cubillos C, Roncagliolo S, Cabrera-Paniagua D. Learning and Motivation When Using Multiple-Try in a Digital Game for Primary Students in Chile. Education Sciences. 2023; 13(11):1119. https://doi.org/10.3390/educsci13111119

Chicago/Turabian Style

Cubillos, Claudio, Silvana Roncagliolo, and Daniel Cabrera-Paniagua. 2023. "Learning and Motivation When Using Multiple-Try in a Digital Game for Primary Students in Chile" Education Sciences 13, no. 11: 1119. https://doi.org/10.3390/educsci13111119

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop