Abstract
Digital technology is considered to have great potential to promote learning in higher education. In line with the Interactive, Constructive, Active, Passive (ICAP) framework, this seems to be particularly true when instructors stimulate high-quality learning activities such as constructive and interactive learning activities instead of active and passive learning activities. Against the background of a lack of empirical studies in authentic, technology-enhanced instructional settings, we investigated the cognitive and affective-motivational effects of these learning activity modes in technology-enhanced higher education courses. To this end, we used 3.820 student assessments regarding 170 course sessions for which the teachers stated the learning activities students were engaged in. Results of multilevel structural equation modelling highlight the importance of technology-enhanced interactive learning activities for students’ perception of learning and the potential negative consequences of passive learning activities for affective-motivational outcomes. However, the superiority of constructive and interactive learning activities compared to passive and active learning activities for cognitive and affective-motivational outcomes was not supported by the findings. Instead, the findings point to potential differential effects of the individual learning activities within one activity mode. Future research should follow up on these effects to gain a more fine-grained understanding of how technology-enhanced learning activities can be optimized to enhance students’ learning outcomes.
Similar content being viewed by others
Introduction
Over 50 years of research have led to the conclusion that the general impact of digital technology on students’ performance in higher education is characterized by a small positive effect at best1,2. Yet, considering more specific relationships, a meta-analysis documented a positive medium effect of digital games on students’ learning achievement in STEM education when compared to traditional instruction3. In contrast, a meta-analysis on the effects of teaching with PowerPoint found no differences in students’ learning outcomes when comparing PowerPoint-based teaching with traditional chalkboard-based teaching4. Taken together, these and other meta-analyses5,6,7 indicate that it is not so much technology per se that determines its effects, but rather how it is used by teachers and students. Accordingly, researchers in the past and present have recommended moving away from so-called media comparisons. Instead, it should be investigated how to best instructionally implement different technologies in classrooms by focusing on the quality of the technology-enhanced learning activities8,9.
To this end, scholars have already proffered suitable conceptual frameworks. Specifically, a prominent instructional model that assumes qualitative differences between four different modes of observable learning activities (interactive, constructive, active, and passive) in terms of students’ associated learning is the ICAP framework10,11,12,13. Despite the ICAP framework becoming more and more established in the field of educational research, there is still a lack of empirical studies that investigate the effects of different learning activity modes on cognitive outcomes, particularly in authentic technology-enhanced higher education settings. Also, numerous theories point to the fact that learning is not just a cognitive, but also an affective-motivational endeavour. However, this aspect has not yet been considered in ICAP research. Against this backdrop, in the present study, we investigated the effects of different technology-enhanced learning activity modes on students’ perceptions of their learning, situational interest and joy in higher education.
Quality of technology implementation based on the ICAP framework
The ICAP framework10,11,12,13 differentiates four observable modes of learning activities that are assumed to be associated with different knowledge change processes, resulting in different cognitive learning outcomes. These four modes of passive, active, constructive, and interactive learning will be outlined in the following:
Learning activities are defined as passive if learners attend towards the learning material and absorb its information without any other observable learning activity. Passive activities might include listening to a podcast or watching a video and are believed to typically only allow isolated storage of information in terms of knowledge-change processes. As a result, learners might only recall the acquired knowledge in identical contexts or when receiving cues for activation. Note that common classroom activities in higher education, such as plenary discussions and presentations, are considered predominantly passive because only a few students will be constructive or interactive at a particular time and most students would only follow the discussion and presentation without doing anything else13.
Active learning includes learning activities in which learners observably and physically manipulate learning materials, including rotating virtual three-dimensional objects, pausing or rewinding online videos, and underlining (important) text passages of a hypermedia text. The active activity mode is considered to encourage the activation of prior knowledge and linking new information and prior knowledge by adapting and complementing schemata. These knowledge construction processes are supposed to facilitate an application of the acquired knowledge in contexts similar to the acquisition context.
Student learning activities are identified as being constructive if students create learning products that go beyond the information that were originally provided by the learning materials. Examples for constructive activities are creating a digital concept map or visualization. A cognitive process that is thought to be typically connected to constructive activities is inferring new knowledge from the activated prior knowledge and the linked knowledge by means of revising, replacing, or restructuring stored knowledge and identifying lacking knowledge. Consequently, a transfer of knowledge should be enabled.
Finally, interactive learning activities take place when students engage in conversations with one or more learning partners characterized by students’ statements being of a constructive nature (instead of passive or active) and a sufficient degree of turn taking. This is supposed to occur when students argue for or against a position in a discussion forum or when students jointly develop a concept for an explanatory video using a visualization tool. Interactive learning activities are assumed to facilitate the acquisition of new knowledge by linking knowledge with the help of learning material (see constructive learning activities) and contributions of the learning partners, in a co-inferring process. In terms of learning outcomes, a joint construction of new (innovative) knowledge should be likely, meaning that students generate knowledge they would not have been able to generate on their own.
Integrating the assumptions on the four learning activity modes, the ICAP hypothesis is as follows: engaging in interactive learning activities should be superior to engaging in constructive learning activities which in turn should be superior to engaging in active learning activities regarding students’ knowledge acquisition. Engaging in passive learning activities is meant to be associated with the lowest outcomes. In addition, Chi and colleagues11 consider the gulf between shallow activities that do not include the process of inferring (meaning passive and active learning activities), and deep activities that do include the process of inferring (meaning constructive and interactive learning activities) to be bigger than within the shallow and deep learning activities. This perspective aligns well with other theoretical models (such as Select-Organize-Integrate14) which differentiate between non-generative and generative learning activities15. As the ICAP hypothesis is supposed to be “applicable to all students, all grade levels, and for all content domains”16 when considering certain prerequisites and limitations11,13, it can provide higher education teachers with (potentially) observable indicators for high-quality technology-enhanced teaching.
Research on ICAP in (technology-enhanced) higher education settings
To test the ICAP hypothesis, first, Chi10 and Chi and Wylie13 presented a reanalysis of around 40 existing studies (not exclusively regarding higher education). Despite seemingly supporting the ICAP hypothesis, this analysis may be criticized due to a positive selection resulting from a non-systematic search and presentation of findings. This circumstance is reflected in the lack of information on the search procedure and is emphasized by Chi and Wylie themselves13. Second, an experimental study18 was performed in which the four learning activity modes were manipulated in a controlled setting. After an individual pretest and reading of an introductory text, all participants were randomly assigned to one of the ICAP conditions, received a fixed time frame to work on the allocated learning activity, and finally took an individual posttest. The results largely confirmed the ICAP hypothesis. However, in less controlled or more authentic learning contexts in which the ICAP hypothesis has been tested so far, the results are more mixed17,18,19,20. Yet, in most of these studies, at least the assumed effects of deep and shallow learning activity modes could be found. For example, students’ constructive/interactive contributions in a MOOC online discussion forum were found to be more conducive to students’ knowledge gain than active contributions20. In a study on the effects of peer instruction (discussion of answers in small groups after answering quiz questions with audience response systems) with a passive, constructive, and interactive condition, only students learning interactively substantially surpassed those who were learning passively21. While these study results are generally in line with the ICAP hypothesis, or at least do not point to a different hierarchy of learning activities, there are several important aspects that previous studies have not addressed yet: first, most studies have not considered authentic technology-enhanced learning contexts. In such contexts students are typically not only asked to perform a single activity to approach a topic (e.g., watching a video) like it is often the case in experimental studies. Instead, they are expected to combine multiple learning activities, move from one to another or switch between them (e.g., watching a video, answering quiz questions, discussing answers with peers, writing a collaborative blog entry on the video). If such contexts were addressed, students’ self-report data on their technology-enhanced learning activities were used only on a coarse-grained level (assessment on technology-enhanced learning activities in all the course at the end of the term21).
Second, few studies have considered multiple learning activities (e.g., listening, reading, watching) within one learning activity mode (e.g., passive) and investigated whether these have comparable effects on knowledge acquisition, as implied by the ICAP model. Though Chi and Wylie13 acknowledge that some learning activities might be more cognitively challenging than others within one mode, the ICAP framework does not make any assumptions about their differential effectiveness.
Third, despite their established importance for students’ learning processes22, students’ motivation (e.g., situational interest) and emotions (e.g., joy) have not been included in ICAP-based studies to date. Regarding the enhancement of students’ motivation as well as positive emotions, theories point to the relevance of learning arrangements in which students experience autonomy and control over a task, consider the task at hand useful and have a sense of belonging23,24,25,26,27. We believe that these processes are more likely to occur when students engage in constructive and interactive learning activities compared to passive and active learning activities. This might be due to the kind of tasks that ask students to generate new, not solely teacher-provided knowledge and therefore engage in (joint) deep processing of the learning content. Empirical findings in support of these assumptions originate from meta-analyses on problem-based learning and computer-supported collaborative learning. In both fields, small to medium positive effects on students’ motivational outcomes could be observed when compared to more traditional and mostly non-collaborative control conditions28,29,30,31. They might be well interpreted based on the ICAP learning modes, as problem-based learning mostly focuses on constructive and interactive learning activities32 and computer-supported collaborative learning mostly focuses on interactive learning activities. Thus, in ICAP terms, this might mean that deep interactive and constructive learning activities should be more conducive to students’ affective-motivational outcomes than shallow passive and active learning activities.
The present research
Based on the ICAP framework and the findings presented above, we investigated the effects of the four different technology-enhanced learning activity modes reported by teachers on students’ cognitive as well as affective-motivational outcomes. We asked higher education teachers of different domains to report on students’ learning activities that occurred in their actual classes while asking students to report on their learning outcomes in end-of-session evaluations. By assessing various course sessions, we aimed to capture authentic learning settings in which a wide range of learning activities might occur over time. We assumed that constructive and interactive learning activities (i.e., “deep” learning activities in Chi and Wylie’s terms) are more positively related to students’ learning (cognitive outcome, H1) and students’ situational interest and joy (affective-motivational learning outcomes, H2) than passive and active (i.e., “shallow”) learning activities. Furthermore, we sought to explore to what degree different effects of single learning activities within one mode on the cognitive and affective-motivational learning outcomes might occur.
Methods
To investigate our hypotheses, we asked 87 university teachers to provide details on the individual learning activities their students engaged in a particular session. Simultaneously, the students evaluated their learning in these sessions using end-of-session paper-and-pencil surveys (students’ perceptions of learning, situational interest, joy). We used the data from all sessions that were responded to by the teachers and evaluated by at least five students. This data is part of a larger project on faculty motivation where we investigated different research questions to the present work (focused on faculty members and their motivations, instead of technology-enhanced learning activities). Specifically, in research questions unrelated to the present investigation, it has been studied to which end faculty motivations are associated with faculty members’ professional learning33, how stable and context specific faculty members’ goals and emotions are34, and how they relate to their basic need satisfaction35 (reported in each session by the faculty members, alongside the assessment of their learning activities). This data was also used in another publication in which these faculty motivations and emotions were compared to analogous assessments during COVID-1936. Further, in past research on this data, we investigated how students’ assessments of teaching quality are linked to faculty members’ teaching motivations37. This last publication contains partial data overlap with the present investigation in that the two of the same dependent variables were considered (students’ evaluations of learning and joy), but with a subsample only. The study was approved by the University of Mannheim IRB board, carried out in accordance with the relevant guidelines and regulations, and participation was voluntary and incentivized through small thank-you gifts provided to the participating teachers. Through e-mails to the department heads, we informed about the study and asked faculty members to participate. Those who agreed to participate received further information to be sent to their students, informing them about the study. Prior to participation, all participants provided informed consent.
Sample
The sample stems from a medium-sized university in the south of Germany, the University of Augsburg. Altogether, our sample consisted of 3.820 student assessments regarding 170 sessions of 42 different courses. The participating students had a mean study duration of 3.5 semesters (SD = 2.1), 81% of them were women. The teachers were on average 42.1 (SD = 11.6) years old, 59.6% of them held a PhD, and about 2 out of 3 were female (68.4%). They were recruited from multiple departments and came from the fields of education sciences, sociology, sports sciences, mathematics, economics, and psychology. At the time of data collection, the University of Augsburg had seven faculties, which are located in different buildings. Most of the included students and teachers were part of the faculty of philosophy and social sciences. We primarily recruited teachers from this faculty due to practical considerations (the study involved research assistants administering the surveys in each session).
Measures
For each session, the teachers made assessments regarding students’ learning activities using a half-structured and open-ended scale (see Supplementary Material S1/2 online). Therein, we first asked the teachers about the total length of the session. This was followed by an instruction in which we detailed what we meant by “types of learning activities” that students performed during the respective sessions (e.g., listen, take notes, summarize content, develop ideas with other students). Likewise, we then illustrated what we meant by “digital technology tools” that the teachers might have used to facilitate these learning activities (presentation software, quiz tools, text editors, video editing software, etc.). Participants were then asked to chronologically list the different learning activities students engaged in during their session, how long these were and whether digital tools were used (e.g., generation of digital mind maps), addressed (e.g., discussion about a shown video), or not used/addressed (e.g., reading of printed text) for each of them. The development of the scale was based on previous studies21,38 in which teachers’ (intended) ICAP-based use of technology was measured. In the present research, we specifically assessed time spent on the individual learning activities to allow for a differentiated measurement of each mentioned activity. Subsequently, using a coding scheme based on Chi and Wylie13 and Chi et al.11 two trained undergraduate research assistants performed a content analysis. The independent coders manually rated the mentioned activities along the ICAP dimensions and classified each mentioned activity as either passive (e.g., watch video), active (e.g., search of information), constructive (e.g., concept mapping), or interactive (e.g., developing group presentation). Data which could not be identified as an activity due to lacking information was classified as other (e.g., “computer”). Following typical recommendations39, we calculated inter-rater reliability based on a random subsample of 10% of the data, which yielded a good reliability of Cohen’s k = 0.88. For the analyses, we then used the duration of each of the four activity modes relative to the overall duration of the session (e.g., if the session was 90 min long, and a total of 45 min were reported for digital, passive learning activities, then the score was 45/90 = 0.5, reflecting that half of the learning time in that session was spent on these respective activities).
The sessions were evaluated by the students regarding their perceptions of learning, situational interest, and joy. For students' perceptions of learning, we used two items of the respective subscale from the SEEQ40,41,42 that assessed students’ evaluation in terms of learning (e.g., “I have learned a lot in today’s session”; Spearman-Brown: r = 0.71). Situational interest was assessed with a single item43 that read as “I found today’s session very interesting”. Finally, joy was assessed with the following single item44: “In today’s session I felt joy”. All items were responded to on a Likert-type scale ranging from 1 (do not agree at all) to 8 (agree completely). Intraclass correlations (student assessments per each individual session: ICC1 = 0.07–0.12; ICC2 = 65–0.75) pointed to substantial shared portions of variance in students’ session-specific assessments that were reliably assessed among them.
Analyses
We conducted multilevel analyses (student evaluations, nested in different course sessions) in which we regressed students’ perceptions of learning, situational interest, and joy on the learning activity modes reported by the teachers regarding the different sessions. Simultaneously, we controlled for teachers’ age and gender (given documented differences based on these aspects regarding student evaluations of teaching quality and teachers’ perceptions and use of digital technology45). The multilevel model and its variables are depicted in the Supplementary Material S3 online. Note that time spent on the four different learning activity modes was not linearly dependent, as teachers also spent time on other activities. We used the model constraint command to identify which regression weights were statistically significantly different from each other. The analyses were conducted in Mplus (Version 8.1) using MLR as an estimator.
Results
Descriptively, we found that teachers reported on average three (M = 2.95, SD = 1.29) different learning activities that students engaged in their sessions. Almost all of them could adequately be classified into one of the four ICAP learning activity modes, only 5% of the mentioned learning activities were not classifiable and placed into the “other” category. Typically, this was due to the teachers not having described the respective learning activity in sufficient detail.
Most of the time, these learning activities were facilitated with digital technologies: 70% of learning time was spent with direct use of digital technology (e.g., generation of digital mind maps), 9% while addressing digital technology (e.g., discussion about a shown video), and 19% without use of digital technology (e.g., reading of printed text). In the following, we focus on the learning activities facilitated with direct use of digital technology. We found that for an average lesson with a time of 80 min, passive learning activities had the highest share with 41.5 (SD = 44.6) minutes, followed by active learning activities with 12.5 (SD = 25.4) minutes, constructive learning activities with 3.3 (SD = 10.5) minutes, and interactive learning activities with 4.5 (SD = 11.7) minutes. Passive learning activities (252 in total) primarily included listening (66), discussion (48), and (student) presentations (41); active learning activities (64 in total) included taking notes (15), textual work (5), and using work sheets (5), constructive activities (32 in total) included summarizing information (6), developing ideas (6), and making a video (4), interactive learning activities (39 in total) primarily consisted of group work (29), but also partner discussions (3) and preparation of group presentations (3).
Associations between teachers’ use of ICAP learning modes with digital technology and students’ learning outcomes (hypothesis 1 and 2)
Results of the multilevel modelling are reported in Table 1. Regarding hypothesis 1, the findings showed that students reported statistically significantly better learning, the more teachers reported to have used interactive digital learning activities in their sessions. Conversely, and regarding hypothesis 2, the more teachers reported using passive learning activities in a session, the less joy reported students to have experienced in the respective session. All other regression weights were not statistically significant. The effect of passive learning time on joy was statistically smaller than the effect of constructive learning time on joy (p < 0.05), all other regression weight comparisons did not reach statistical significance.
Exploratory analyses on differences between different passive learning activities
In the second model, we followed up on the passive learning activities that were largely reported by the faculty members. We regressed the same dependent variables on the three most prevalent types of passive learning activities: listening, plenary discussion, and (student) presentation. The results showed that students reported higher situational interest, the longer learning with presentations was reported by the teachers. Further, students reported experiencing less joy, the longer teachers reported to evoke passive listening activities. The other effects were not statistically significant. Comparisons of the regression weights confirmed that the effects of presentations on students’ perceptions of learning and situational interest were statistically larger than those of listening and plenary discussions on their perceptions of learning and situational interest.
Discussion
To evaluate higher education teachers’ technology implementation not only from a perspective of quantity but also quality, ICAP holds the potential to serve as a valuable framework. However, several questions regarding its application in authentic (technology-enhanced) learning contexts remain open. With the present research, we tried to remedy some of these issues by investigating to what degree the ICAP hypothesis holds true for cognitive as well as affective-motivational outcomes, in authentic technology-enhanced higher education course sessions, and to what degree there might be differences in the effects of single technology-enhanced learning activities within one particular learning activity mode.
Descriptively, the associations between the four learning activity modes, students' perceptions of learning, situational interest as well as joy were roughly in line with the ICAP hypothesis13, findings on problem-based learning and computer-supported collaborative learning28,29,30,31,32. However, only technology-enhanced interactive learning activities could explain students’ more positive perceptions of their learning, whereas engagement in the other learning activity modes did not seem to be substantially aligned with students’ perceptions of learning. Regarding affective-motivational outcomes, only passive learning activities were negatively associated with joy. We could only observe a significant difference between the shallow passive and the deep constructive learning activity mode for joy in favour of the constructive activity mode. Consequently, the pattern of results provides empirical support for the “end points” of the ICAP-continuum, i.e., passive and interactive learning activity modes, and at the very least does not contradict the generally assumed hierarchy of learning activity modes. Yet, we could not confirm our hypotheses. We likely have to consider different factors to explain the lacking effects.
First, the present study differs from previous ICAP-based studies in that it was implemented in authentic learning settings. In such settings, different learning activities are typically combined or arranged in sequences. Therefore, the execution of a certain learning activity might enhance or impair the execution of a subsequent activity. For example, in two studies students showed better cognitive skills when receiving a modelling example (passive instruction) before problem-solving (constructive instruction) in an alternate manner compared to when being confronted with problem-solving tasks prior to modelling examples, particularly when students had low prior knowledge46. Thus, in line with Roscoe et al.47, we suggest further investigations on the sequence of learning activities to determine if activity order, e.g., from low to high engagement or vice versa, influences students' efforts and learning outcomes.
Second, teachers might tend to report students’ learning activities according to their intended student behaviours instead of the kind of learning they actually encourage by their instruction or their perceived student behaviours. The following observation confirms this explanation: Despite their initial intentions (e.g., designing a constructive or interactive lesson), teachers were inclined to use directives and provide questions in activity sheets that might initiate active learning activities (e.g., label, match11). Consequently, even though teachers’ assessment of students’ activities might be more accurate than students’ own assessment21, teachers’ bias should be still accounted for in future research (by introducing ICAP-oriented lesson transcript-based assessments of lesson tasks48).
Third, the exploratory analysis of the passive learning activities revealed that single learning activities within one mode can produce different effects. In concrete terms, while (student) presentations produced positive effects on situational interest, listening (to teachers) produced negative effects on students’ joy. In addition, presentations were more conducive to students’ perceptions of learning and situational interest than listening and plenary discussion. These differences within one mode might have resulted in the lacking effects of the learning activities by cancelling each other's positive and negative effects out. Consequently, the results raise the question if the ICAP learning activity modes might be too coarse-grained to capture learning activities and their effects accurately.
Despite these study-specific explanations, the theoretical assumptions stated in the ICAP framework might itself impose limitations on the study of learning activities. Though there might be an association between directly observable behaviour and cognitive processes, it is rather a matter of probability than a deterministic relation13. However, elaboration and other deep-level learning processes are mainly responsible for students’ learning gains.
Irrespective of the strengths of our study (including the authentic learning contexts, the pairing of student and teacher reports, and the fine differentiation of learning activities), methodological aspects also must be considered that might have influenced our findings: First, the teachers in our sample tended to only make generic notes on students’ learning activities which could have resulted in a too inclusive or exclusive coding and lacking differences between the activities. Second, students’ perceptions of learning as measured in our study focused on the quantity of learning (e.g., “I have learned a lot in today’s session”). In other words, we might not have been able to capture the quality of learning in terms of deep knowledge-change processes assumed to take place when engaging in deep learning activities. Third, and related, the study only relies on self-reported (albeit stemming from both teachers and students) data which might have resulted in an assessment bias. Fourth, our sample contained data regarding courses from multiple subjects, primarily from the humanities and social sciences. This could limit generalizability to other fields (e.g., medical education), and there might be subject-specific effects that could be informative for future research to follow up on. Last, due to the low frequency of reported single learning activities within the active, constructive and interactive learning activity mode, we could only compare single learning activities within one mode but not across different modes.
In conclusion, the ICAP framework is considered helpful as a theoretical model to describe and classify technology-enhanced learning activities in terms of the quality of technology implementation. It might assist higher education teachers to plan, implement, and evaluate their technology-enhanced teaching. Our results particularly establish the importance of technology-enhanced interactive learning activities for students’ learning success as well as the potential negative consequences of technology-enhanced passive learning activities for students’ joy. Thus, higher education teachers should be supported in using technology in their teaching in a more interactive and less passive way. However, regarding future empirical research, further details in the ICAP framework might need to be amended. Based on our results, we suggest following up on the effects of combining or sequencing different learning activities and of single learning activities within one mode.
Data availability
The dataset and the analytical code underlying the presented analyses are available from the corresponding author on reasonable request. We will publish the full dataset after an embargo period.
References
Tamim, R. M., Bernard, R. M., Borokhovski, E., Abrami, P. C. & Schmid, R. F. What forty years of research says about the impact of technology on learning: a second-order meta-analysis and validation study. Rev. Educ. Res. 81(1), 4–28. https://doi.org/10.3102/0034654310393361 (2011).
Sailer, M., Maier, R., Berger, S., Kastorff, T. & Stegmann, K. Learning activities in technology-enhanced learning: a systematic review of meta-analyses and second-order meta-analysis in higher education. Learn. Individ. Differ. 112, 102446. https://doi.org/10.1016/j.lindif.2024.102446 (2024).
Wang, L. et al. Effects of digital game-based STEM education on students’ learning achievement: a meta-analysis. Int. J. STEM Educ. https://doi.org/10.1186/s40594-022-00344-0 (2022).
Baker, J. P., Goodboy, A. K., Bowman, N. D. & Wright, A. A. Does teaching with PowerPoint increase students’ learning? A meta-analysis. Comput. Educ. 126, 376–387. https://doi.org/10.1016/j.compedu.2018.08.003 (2018).
Chien, Y.-T., Chang, Y.-H. & Chang, C.-Y. Do we click in the right way? A meta-analytic review of clicker-integrated instruction. Educ. Res. Rev. 17, 1–18. https://doi.org/10.1016/j.edurev.2015.10.003 (2016).
Hunsu, J., Adesope, O. & Bayly, J. D. A meta-analysis of the effects of the audience response systems (clicker-based technologies) on cognition and affect. Comput. Educ. 94, 102–119. https://doi.org/10.1016/j.compedu.2015.11.013 (2016).
Schmidt, R. F. et al. The effects of technology use in postsecondary education: a meta-analysis of classroom applications. Comput. Educ. 72, 271–291. https://doi.org/10.1016/j.compedu.2013.11.002 (2014).
Clark, R. E. Media will never influence learning. Educ. Tech. Res. Dev. 42(2), 21–29. https://doi.org/10.1007/BF02299088 (1994).
Fütterer, T., Scheiter, K., Cheng, X. & Stürmer, K. Quality beats frequency? Investigating students’ effort in learning when introducing technology in classrooms. Contemp. Educ. Psychol. 69, 102042. https://doi.org/10.1016/j.cedpsych.2022.102042 (2022).
Chi, M. T. Active-constructive-interactive: a conceptual framework for differentiating learning activities. Top. Cogn. Sci. 1(1), 73–105. https://doi.org/10.1111/j.1756-8765.2008.01005.x (2009).
Chi, M. T. H. et al. Translating the ICAP theory of cognitive engagement into practice. Cognit. Sci. 42(6), 1777–1832. https://doi.org/10.1111/cogs.12626 (2018).
Chi, M. T. H. & Boucher, N. S. Applying the ICAP framework to improve classroom learning. In In Their Own Words. What Scholars and Teachers Want You to Know About Why and How to Apply the Science of Learning in Your Academic Setting (eds. Overson, C. E., Hakala, C. M., Kordonowy, L. L. & Benassi, V. A.). 93–110. https://teachpsych.org/ebooks/itow (Society for the Teaching of Psychology, 2023).
Chi, M. T. & Wylie, R. The ICAP framework: linking cognitive engagement to active learning outcomes. Educ. Psychol. 49(9), 219–243. https://doi.org/10.1080/00461520.2014.965823 (2014).
Mayer, R. E. Cognitive theory of multimedia learning. In The Cambridge Handbook of Multimedia Learning (ed. Mayer, R. E.). 43–71. https://doi.org/10.1017/CBO9781139547369.005 (Cambridge University Press, 2014).
Fiorella, L. & Mayer, R. E. Eight ways to promote generative learning. Educ. Psychol. Rev. 28(4), 717–741. https://doi.org/10.1007/s10648-015-9348-9 (2016).
Chi, M. T. H. Translating a theory of active learning: an attempt to close the research-practice gap in education. Top. Cognit. Sci. 13, 441–463. https://doi.org/10.1111/tops.12539 (2021).
Menekse, M., Stump, G. S., Krause, S. & Chi, M. T. Differentiated overt learning activities for effective instruction in engineering classrooms. J. Eng. Educ. 102(3), 346–374. https://doi.org/10.1002/jee.20021 (2013).
Henderson, J. B. Beyond, “active learning”: how the ICAP framework permits more acute examination of the popular peer instruction pedagogy. Harvard Educ. Rev. 89(4), 611–634. https://doi.org/10.17763/1943-5045-89.4.611 (2019).
Margulieux, L. E. & Catrambone, R. Finding the best types of guidance for constructing self-explanations of subgoals in programming. J. Learn. Sci. 28(1), 108–151. https://doi.org/10.1080/10508406.2018.1491852 (2019).
Wang, X., Wen, M. & Rose, C. P. Towards triggering higher-order thinking behaviors in MOOCs. In LAK ’16 Conference Proceedings: The Sixth International Conference on Learning Analytics & Knowledge (eds. Gasevic, D. et al.). 398–407. https://doi.org/10.1145/2883851.2883964 (2016).
Wekerle, C., Daumiller, M. & Kollar, I. Using digital technology to promote higher education learning: the importance of different learning activities and their relations to learning outcomes. J. Res. Technol. Educ. 54, 1–17. https://doi.org/10.1080/15391523.2020.1799455 (2020).
Linnenbrink-Garcia, L., Patall, E. A. & Pekrun, R. Adaptive motivation and emotion in education: research and principles for instructional design. Policy Insights Behav. Brain Sci. 3(2), 228–236. https://doi.org/10.1177/2372732216644450 (2016).
Ryan, R. M. & Deci, E. L. Self-determination theory and the facilitation of intrinsic motivation, social development, and well-being. Am. Psychol. 55(1), 68–78. https://doi.org/10.1037/0003-066X.55.1.68 (2000).
Frenzel, A. C., Pekrun, R. & Goetz, T. Perceived learning environment and students´ emotional experiences: a multilevel analysis of mathematics classrooms. Learn. Instruct. 17(5), 478–493. https://doi.org/10.1016/j.learninstruc.2007.09.001 (2007).
Eccles, J. S. & Wigfield, A. Motivational beliefs, values and goals. Annu. Rev. Psychol. 53, 109–132. https://doi.org/10.1146/annurev.psych.53.100901.135153 (2002).
Pekrun, R. A social-cognitive, control-value theory of achievement emotions. In Motivational Psychology of Human Development: Developing Motivation and Motivating Development (ed. Heckhausen, J.). 143–163. https://doi.org/10.1016/S0166-4115(00)80010-2 (Elsevier, 2002).
Pintrich, P. R. A motivational science perspective on the role of student motivation in learning and teaching contexts. J. Educ. Psychol. 95(4), 667–686. https://doi.org/10.1037/0022-0663.95.4.667 (2003).
Batdi, V. The effects of a problem based learning approach on students attitude levels: a meta-analysis. Educ. Res. Rev. 9(9), 272–276. https://doi.org/10.5897/ERR2014.1771 (2014).
Demirel, M. & Dağyar, M. Effects of problem-based learning on attitude: a meta-analysis study. Eur. J. Math. Sci. Technol. Educ. 12(8), 2115–2137. https://doi.org/10.12973/eurasia.2016.1293a (2016).
Sung, Y., Yang, J. & Han-Yueh, L. The effects of mobile-computer-supported collaborative learning: meta-analysis and critical synthesis. Rev. Educ. Res. 87(4), 768–805. https://doi.org/10.3102/0034654317704307 (2017).
Wijnia, L. et al. The effects of problem-based, project-based, and case-based learning on students' motivation: a meta-analysis. https://www.nro.nl/sites/nro/files/migrate/Wijnia_Eindrapport_NRO.pdf (2017).
Hmelo-Silver, C. E. Problem-based learning: what and how do students learn?. Educ. Psychol. Rev. 16(3), 235–266. https://doi.org/10.1023/B:EDPR.0000034022.16470.f3 (2004).
Hein, J., Janke, S., Daumiller, M., Dresel, M. & Dickhäuser, O. No learning without autonomy? Moderators of the association between university instructors’ learning goals and learning time in the teaching-related learning process. Learn. Individ. Differ. 83–84, 01937. https://doi.org/10.1016/j.lindif.2020.101937 (2020).
Daumiller, M. & Dresel, M. Temporal dynamics between faculty goals, burnout/engagement, and performance in teaching and research: a latent change score approach. Contemp. Educ. Psychol. 72, 102124. https://doi.org/10.1016/j.cedpsych.2022.102124 (2023).
Keller, M. V. et al. Intertwining self-efficacy, basic psychological need satisfaction, and emotions in higher education teaching: a micro-longitudinal study. Soc. Psychol. Educ. https://doi.org/10.1007/s11218-024-09888-1 (2024).
Schwab, C. et al. “I’m tired of black boxes!”: A systematic comparison of faculty well-being and need satisfaction before and during the COVID-19 crisis. PLoS ONE 17(10), e0272738. https://doi.org/10.1371/journal.pone.0272738 (2022).
Daumiller, M. et al. Do teachers’ achievement goals and self-efficacy beliefs matter for students’ learning experiences? Evidence from two studies on perceived teaching quality and emotional experiences. Learn. Instruct. 76, 101458. https://doi.org/10.1016/j.learninstruc.2021.101458 (2021).
Wekerle, C. & Kollar, I. Using technology to promote student learning? An analysis of pre- and in-service teachers’ lesson plans. Technol. Pedagogy Educ. 31(5), 597–614. https://doi.org/10.1080/1475939X.2022.2083669 (2022).
Neuendorf, K. A. The content analysis guidebook. (Sage, 2017).
Marsh, H. SEEQ: a reliable, valid, and useful instrument for collecting students’ evaluations of university teaching. Br. J. Educ. Psychol. 52(1), 77–95. https://doi.org/10.1111/j.2044-8279.1982.tb02505.x (1982).
Daumiller, M. et al. Do teachers’ achievement goals and self-efficacy beliefs matter for students’ learning experiences? Evidence from two studies on perceived teaching quality and emotional experiences. Learn. Instruct. 76, 76101458. https://doi.org/10.1016/j.learninstruc.2021.101458 (2021).
Daumiller, M., Grassinger, R., Engelschalk, T. & Dresel, M. SEEQ-DE. Diagnostica https://doi.org/10.1026/0012-1924/a000274 (2021).
Wilde, M., Bätz, K., Kovaleva, A. & Urhahne, D. Überprüfung einer Kurzskala intrinsischer Motivation (KIM) [Review of a short scale of intrinsic motivation (KIM)]. Z. Didaktik Naturwiss. 15 (2009).
Götz, T., Sticca, F., Pekrun, R., Murayama, K. & Elliot, A. J. Intraindividual relations between achievement goals and discrete achievement emotions: an experience sampling approach. Learn. Instruct. 41, 115–125. https://doi.org/10.1016/j.learninstruc.2015.10.007 (2016).
Fraillon, J. et al. Preparing for life in a digital world: IEA International Computer and Information Literacy Study 2018. Int. Rep. https://doi.org/10.1007/978-3-030-38781-5(Springer (2020).
Murböck, J. The order matters: sequencing strategies in example-based learning. Dissertation, Ludwig-Maximilians-Universität München. https://doi.org/10.5282/edoc.22989 (2018).
Roscoe, R. D., Gutierrez, P. J., Wylie, R. & Chi, M. T. H. Evaluation lesson design and implementation within the ICAP framework. In Learning and Becoming in Practice: The International Conference of the Learning Sciences (ICLS) 2014. Vol. 2 (eds. Polman, J. L. et al.). 972–976. https://repository.isls.org//handle/1/1223 (International Society of the Learning Sciences, 2014).
Vosniadou, S. et al. Using an extended ICAP-based coding guide as a framework for the analysis of classroom observations. Teach. Teacher Educ. 128, 104133. https://doi.org/10.1016/j.tate.2023.104133 (2023).
Acknowledgements
We would like to thank Sarah Hell and Jeannine Schwarz for their support in coding the data.
Funding
Open Access funding enabled and organized by Projekt DEAL. This study was funded by the German Research Foundation (Grants DA 2392/2-2, DI 929/5-2, DR 454/8-2, JA 3137/2-2) as well as Stiftung Innovation in der Hochschullehre (Grant FBM2020-EA-2620-01350).
Author information
Authors and Affiliations
Contributions
All authors conceptualised the study. Both first authors analysed the data, and wrote and revised the main manuscript text. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Wekerle, C., Daumiller, M., Janke, S. et al. Putting ICAP to the test: how technology-enhanced learning activities are related to cognitive and affective-motivational learning outcomes in higher education. Sci Rep 14, 16295 (2024). https://doi.org/10.1038/s41598-024-66069-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-66069-y