Optimizing learning in undergraduate psychology students: the impact of advance quizzing, review, and classroom attendance
© The Author(s) 2017
Received: 30 December 2016
Accepted: 20 August 2017
Published: 20 September 2017
This study investigates whether introducing simple cognitive interventions that are known to enhance learning in laboratory studies can be transferred to classroom settings. In an introductory psychology class, students were provided with a brief advance quiz on the topic covered in each particular lecture. In case they did not attend class, they still had the opportunity to fill it out at home. The interventions were offered on a voluntary basis and the students were given the opportunity to obtain an extra point towards the final exam by reviewing the key points of each lecture on a regular basis. The results indicate that both advance quizzing and review enhanced the grades on the final examination. Although it is likely that individual differences also contribute to learning outcomes, these effects were specific for the particular course and independent from class attendance. The results indicate that transfer of simple cognitive interventions from laboratory to classroom settings is feasible. They also indicate that these interventions can be effectively implemented into the lectures to boost student learning.
KeywordsMemory Testing effects Pretesting effects Podcasts
Undergraduate university students have only a fragmentary knowledge about effective learning strategies. In contrast, cognitive psychologists have a very fine-grained understanding about strategies and activities that benefit learning in laboratory studies. One of the most important insights is the effect of retrieval as an effective learning strategy. Retrieval benefits learning even for materials to be studied but not yet studied (i.e., advance quizzing). Retrieval induced as a review of key points between lectures also benefits learning. Transferring these strategies and activities into the classroom is an important mission for contemporary cognitive psychology. Although the effects of quizzing and review are small, they are consistent and specific to the learning materials.
Undergraduate university students have only a fragmentary knowledge about effective learning strategies (Bjork, Dunlosky, & Kornell, 2013; Karpicke, Butler, & Roediger, 2009; Putnam, Sunkhasattee, & Roediger, 2016). In contrast, cognitive psychologists have a fine-grained understanding about strategies and activities that benefit learning in laboratory studies. For example, there is convincing evidence that testing is a very effective learning strategy (Roediger & Karpicke, 2006). In fact, even unsuccessful retrieval attempts can enhance subsequent learning (Kornell, Hays, & Bjork, 2009; Richland, Kornell, & Kao, 2009). In order to test whether these findings transfer to undergraduate university student learning in a classroom setting, the present study used a brief quiz at the beginning of each lecture on the topic to be covered in that particular lecture. Beside this intervention, which focused on retrieval before the lecture, a second intervention was administered that focused on retrieval after the lecture. Specifically, the students were given the opportunity to obtain an extra point towards the final exam by reviewing the key points of each lecture on a regular basis.
There is already some evidence that quizzing as a learning tool can be successfully used in the classroom. For example, regularly answering a set of questions that required the retrieval of information from the same day’s class significantly improved subsequent exam performance in a statistics course (Lyle & Crawford, 2011). In an introductory psychology class, Pennebaker, Gosling, and Ferrell (2013) presented a brief computerized quiz using an online system to test for the content of the previous lesson. Their results showed better performance and moreover, a reduction in the gap between students from different social classes. Using an experimental approach in the classroom, Roediger, Agarwal, McDaniel, and McDermott (2011) showed that quizzing had a beneficial effect for middle school students in a social studies course. Similarly, McDaniel, Agarwal, Huelser, McDermott, and Roediger (2011) demonstrated that quizzing improved learning in an eighth-grade science class. In one experiment, they specifically manipulated the placement of quizzing, with some students being tested before the respective lecture was introduced in class. The results showed that, overall, review quizzing produced the greatest increase in final exam performance, advance quizzing also seemed to provide a benefit compared to not having a quiz at all. In the present study, the impact of advance quizzing is addressed in university students, notably without a strict requirement for students to complete every activity.
Although it may seem counterintuitive to answer questions about topics that have not been studied yet, answering questions beforehand activates related knowledge about the topic and makes it easier to connect new information to what is already known. Moreover, it can raise the interest in the topic and thus enhance attention during the lecture (Richland et al., 2009). Thus, it can have both direct and indirect effects. In the present study, four questions were prepared for each lecture on the university online learning platform. This allowed giving individual feedback at the end of the quiz. As classroom attendance was facultative, a podcast of the lecture that contained the link to the advance quiz was made available on the learning platform. By considering the access time to the quiz, it was possible to identify students who attended the class, students who used the online course materials, and those who did neither. Therefore, the study can also contribute to the question whether attending class and/or the availability of recorded lectures is associated with performance on the final exam (cf. Credé, Roch, & Kieszczynska, 2010; Bos, Groeneveld, van Bruggen, & Brand-Gruwel, 2016; Traphagan, Kucsera, & Kishi, 2010).
The study further investigated whether reviewing the material after the lecture would benefit learning. Specifically, students were motivated to review each lecture and to submit the three key points to the learning platform before the next lecture. For the regular submission, they received a bonus point towards the final exam. The idea behind this intervention was to provide the students with retrieval practice and thus to facilitate learning (cf. Roediger & Karpicke, 2006). Moreover, reviewing after a delay can also be considered as spaced practice, which is also known to facilitate learning (Cepeda, Pashler, Vul, Wixted, & Rohrer, 2006). As sleep enhances memory consolidation, the time window for submission did not open until the day after the lecture. Moreover, in line with the reconsolidation hypothesis, which proposes that when a memory is reactivated, the trace may return to a labile state after which time it is reconsolidated or restabilized, the idea was that the requirement for reconsolidation would further strengthen learning (Feld & Diekelmann, 2015).
The interventions were embedded in an introductory psychology course (i.e., “Introduction to Memory”) that students had to attend as part of the curriculum. Due to a change in the curriculum, a larger number of students – from the first and from the second year – were enrolled in the course (i.e., two study years). For the first-year students, the curriculum change involved a new testing format in which the exam was combined with another cognitive psychology introduction course (“Introduction to Perception”) which had taken place one semester earlier. For the benefit of the present study, the performance achieved in this other part of the exam could be used to test the specificity of the intervention effects.
Settings and participants
Participants were undergraduate psychology students from the University of Bern who took the course “Introduction to Memory” as part of their curriculum. The study was conducted in agreement with the ethical guidelines of the Human Science Faculty of the University of Bern. The original number of students enrolled in the course was 673; a total of 300 first-year students and 210 second-year students took the final exam and were included in the analyses. Due to the change in the curriculum two student cohorts (i.e., first-year and second-year students) were taught together. Related to this change, the exam of the first-year students was combined with another course (“Introduction to Perception”), which had taken place one semester earlier. The interval between the last class of that course and the final exam was 26 weeks. For first-year students, performance achieved in this exam was used to test the specificity of the advance quizzing and review interventions. Consequently, the data on first- and second-year students were analysed separately.
Measures and materials
The “Introduction to Memory” course consisted of thirteen lectures distributed across 13 weeks with one lecture per week, and was, to a large part, based on the book by Baddeley, Eysenck, and Anderson (2015). Each lecture was recorded and made available as a podcast on the online learning platform of the university (ilias.unibe.ch). Moreover, the slides for each lecture were uploaded before each lecture, in accordance with the recommendation by Putnam et al. (2016).
For each lecture, four multiple choice questions were prepared. These were accessible via the online learning platform. The questions covered some of the content of the actual lecture and provided immediate feedback about the correct solution (cf. Richland et al., 2009; Butler & Roediger, 2008). These were arranged in two different access windows. The first one was open during the lecture and was used as a proxy to assess classroom attendance. In the first part of each lecture, students were provided with about 5 minutes to answer these questions in class. The second window opened after the lecture and thus the quiz was also available for those students who did not attend the lecture, but studied the particular topic using the online materials (e.g., slides, podcasts etc.). For the purpose of this study, the number of completed quizzes was used. Thus, it was possible to have a maximum score of 13 both for the lecture time window (which was used as a proxy for classroom attendance) and for the time window that opened only after the lecture (which was used as a proxy for podcast use).
In order to motivate the students to review the content of a lecture before the next one, students were given the opportunity to submit three key points via the online learning platform. In order to benefit from overnight consolidation (cf. Feld & Diekelmann, 2015), the opportunity to submit the review opened the day after the lecture and lasted until the day before the next lecture. Thus, it was possible to have a maximum of 13 completed reviews. An extra point towards the final exam was given when at least 12 reviews were submitted via the online platform. This additional point was considered to calculate the final grade. However, for the purpose of this study, it was not added to the points in the final exam.
The final exam took place 3 weeks after the last lecture. It consisted of 22 multiple choice questions, which covered the content of all the lectures (i.e., 1–2 questions per topic). For the first-year students the exam was combined with the exam on the course “Introduction to Perception”, which was prepared accordingly. For each exam a maximum of 22 points was possible. As noted above, the extra point that could be gained by submitting the reviews on a regular basis was not considered to calculate the points for the final exam for the purpose of this study.
In the results section, first the descriptive statistics for each variable are presented. Next, a group comparison is provided to test for differences between first-year and second-year students. The main focus is on the relationship between advance quizzing, review and performance in the final exam, and correlations are provided separately for first-year and second-year students. To test the predictive value of these interventions, separate regression analyses were conducted. For each of these analyses, diagnostic statistics were run to check whether the assumptions were met. The results of the diagnostics are summarized here. Analyses of standard residuals showed that the data contained no outliers (standard residuals < 3). To test whether the data met the assumption of collinearity, VIF scores were calculated, with values between 1.1 and 1.5, indicating that multicollinearity was not a concern. The data also met the assumption of independent errors (with Durbin-Watson values = 2 +/− 0.5). Histograms of standardized residuals and the plots of standardized residuals showed that the errors were approximately normally distributed. The scatterplot of standardized residuals showed that the data also met the assumptions of homogeneity of variance and linearity. The data also met the assumption of non-zero variances. In a set of follow-up analyses, the exclusivity of the contribution of advance quizzing was assessed using a stricter criterion. Using hierarchical regression analyses, the extra contribution of advance quizzing in the classroom was tested after controlling for the influence of quizzing after the lecture and review.
Descriptive statistics for advance quizzing, review and memory exam
Correlation between advance quizzing, review and exam scores in first-year students (normal font), and second-year students (italics), respectively
To test the predictive value of these interventions, three separate regression analyses were conducted. The first analysis focused on the first-year students. The second analysis focused on the second-year students to establish the generality of the results. The third analysis focused on the specificity of the influences of advance quizzing and review by testing whether the interventions would also predict the performance of the first-year students in the “Introduction to Perception” course.
Regression analysis predicting “Introduction to Memory” exam performance from advance quizzing and review in first-year students
Regression analysis predicting “Introduction to Memory” exam performance from advance quizzing and review in second-year students
Regression analysis predicting “Introduction to Perception” exam performance from advance quizzing and review of “introduction to memory” in first-year students
If one considers classroom attendance as a stable trait of a person (e.g., conscientiousness) and if only classroom attendance would predict exam sores, one would expect the same predictive power of advance quizzing in the classroom variable for both the memory and the perception exam. However, a comparison of the standardized beta coefficients shows that the predictive power of this variable is higher for the memory exam, that is, .177 (Table 3) and .180 (Table 4) for first-year and second-year students, respectively, compared to .134 for the perception exam (Table 5). This indicates that not only classroom attendance contributes to this relationship, but also that advance quizzing in the classroom provides additional explanatory power.
The previous analyses evaluated the impact of the interventions assuming that the advance quizzes were always used before learning. However, it is possible that the students who did not attend class may have completed the quizzes after they studied. If so, then the quizzes would not be advance quizzes for those students. It is also possible that students accessed the quizzes during class time and used them as study guides during the lecture rather than using them at the beginning of the lecture as instructed. Again, if so, the quizzes would not be advance quizzes for those students. In order to control for these possibilities further analyses were conducted. In these analyses, the focus was on those students who completed the advance quizzes during class, and specifically, during the first part of the lecture as instructed. The latter was possible due to the availability of the exact time stamp of the access to the advance quiz. To allow for a proper measurement of the impact of advance quizzing at the beginning of the lecture, instances in which students also accessed the advance quiz after the lecture were excluded. As no time stamp information was available for the access to the podcast it was not possible to apply the same restrictions to those students who did not attend class. With these restrictions, the mean number of uses of advance quizzing during the lecture was 3.91 (SD = 3.65) for first-year students and 2.79 (SD = 3.60) for second-year students, which was statistically significant, t(508) = 3.44, p < .01.
To test whether advance quizzing during the lecture had an exclusive predictive value for the final exam, hierarchical regression analyses were conducted that included quizzing after the lecture and review in the first step and advance quizzing at the beginning of the lecture in the second step, separately for first-year students and second-year students.
Hierarchical regression analysis to assess the exclusive contribution of advance quizzing before learning to the prediction of “Introduction to Memory” exam performance in first-year students
Hierarchical regression analysis to assess the exclusive contribution of advance quizzing before learning to the prediction of “Introduction to Memory” exam performance in second-year students
The purpose of the present study was to investigate whether retrieval practice before learning (i.e., advance quizzing) and retrieval practice after learning (i.e., review between lectures) would affect final exam grades in an introductory psychology class. The results indicated that both variables enhanced final grades. Although the effects were small, they were consistent and replicated in two separate student groups (i.e., first-year and second-year students). Moreover, the effect was specific as it was not present in another cognitive psychology introduction course attended by the same first-year students.
These results are in line with other studies that showed beneficial effects of quizzing, retrieval practice, and spaced retrieval (e.g., McDaniel et al., 2011; Richland et al., 2009; Roediger & Karpicke, 2006). It is likely that these interventions have both direct and indirect effects on final exam performance. The specific processing of study materials can lead to greater familiarity and reduce anxiety in particular for complex materials. Quizzing and retrieval can lead to more elaborate associations and retrieval cues, can enhance consolidation, and lead to more elaborate memory traces through reconsolidation. Repeated processing of the materials can also lead to retrieval-induced facilitation for other materials and enhanced transfer of learning (Chan, McDermott, & Roediger, 2006; Rohrer, Taylor, & Sholar, 2010). Advance quizzing may specifically help in building up new memory representations. It may function as an advance organizer that guides attention through the lectures and thus enhances attentiveness. Importantly, due to the voluntary nature of the interventions in this study, self-initiated processes may be particularly boosted. The interventions can also help to optimize students’ learning strategies, e.g., helping students identify what they do not understand, which in turn can trigger further studying, and guide them to schedule their study time (Bjork et al., 2013; Hartwig & Dunlosky, 2012; Karpicke et al., 2009; Putnam et al., 2016).
Overall, the results of the current study indicate that even rather simple interventions based on the insights from basic research in laboratory cognitive psychology can have a significant impact on student learning in the classroom. The results confirm theoretical considerations from laboratory studies on the effects of testing, pretesting, and the forward effect of testing (e.g., Roediger & Karpicke, 2006, Kornell et al., 2009; Richland et al., 2009; Pastötter & Bäuml, 2014). In contrast to previous work, in the present study, the interventions were conceptualized on a voluntary basis, that is, students were not strictly required to participate in the interventions. Thus, the quality of the submitted responses, in particular for the review intervention, may have varied substantially. Nevertheless, participating on a regular basis still improved overall course performance.
The results also shed light on whether or not classroom attendance is necessary for successful performance. They indicate that classroom presence is not a mandatory precondition for success in the final exam. Rather it seems that using the provided materials on the online platform such as the podcast and slides can compensate for the lack of “live” experience. These results are in line with other studies that have found no detrimental effects of new technologies compared to traditional face-to-face lectures (e.g., Bos et al., 2016; Grabe & Christopherson, 2008). They are also encouraging for the growing field of distance education where classroom attendance is not possible at all.
However, a limitation of the present study is the measurement of classroom attendance. It is important to note that advance quizzing in the classroom was used as a proxy for classroom attendance. As some students did not bring their computers/smartphones to class, this measure is not comprehensive. Similarly, using advance quizzing after the class as a proxy for podcast use is also imprecise. In fact, the specific use of podcasts was not assessed at all. Nevertheless, the results indicate that advance quizzing during class and advance quizzing after class both contributed to predicting the final exam performance.
On a related note, it is possible that students who did not attend class may have completed the quizzes after they studied. If so, then the quizzes would not be advance quizzes for those students. Similarly, it is possible that students accessed the quizzes during class time and used them as study guides during the lecture rather than using them before learning. If so, the quizzes would not be advance quizzes for those students. In order to control for these possibilities further analyses were conducted to test for the independent contribution of advance quizzing. In these analyses, the focus was on those students who completed the advance quizzes during class, and specifically, during the first part of the lecture as instructed. In addition, instances in which students also accessed the advance quiz after the lecture were excluded. The results of these analyses showed that advance quizzing before learning contributed exclusively to performance in the final exam over and above the other two variables.
A caveat may also apply to the review intervention. As the specific review submissions were not controlled for quality or accuracy, they also represent a rather vague variable. It is very possible that the predictive value of review would have been even stronger if such a quality control had been implemented. This may be an avenue for future research. Despite these shortcomings, the presence of a consistent enhancing effect is striking.
Finally, due to the correlational approach, it is likely that extraneous variables such as individual differences also influenced the outcome. Specifically, more conscientious students may study more, may be more likely to participate in the offer of learning aids, and may attend class more often. Although the correlation between advance quizzing and final exam performance was much stronger for the “Introduction to Memory” class than for the “Introduction to Perception” class, the latter correlation was still significant. This supports the hypothesis that individual differences also contribute to the learning outcomes.
To conclude, this study demonstrates that simple cognitive interventions transfer from laboratory to classroom settings. Retrieval benefits learning even for materials to be studied but not yet studied (i.e., advance quizzing). Retrieval induced as a review of key points between lectures also benefits learning. Transferring these strategies and activities into the classroom is an important mission for contemporary cognitive psychology. The results indicate that some effective interventions can be implemented rather easily into the lectures to boost student learning.
I would like to thank Stefan Walter for technical support and Janek Lobmaier for providing access to the perception exam scores.
This project did not have a funding source.
Availability of data and materials
The data and materials are made available in Additional file 1.
Ethics approval and consent to participate
The study was conducted in agreement with the ethical guidelines of the Human Science Faculty of the University of Bern.
Consent for publication
The author declares that he has no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
- Baddeley, A. D., Eysenck, M. W., & Anderson, M. C. (2015). Memory. NY: Psychology Press.Google Scholar
- Bjork, R. A., Dunlosky, J., & Kornell, N. (2013). Self-regulated learning: beliefs, techniques, and illusions. Annual Review of Psychology, 64, 417–444. doi:10.1146/annurevpsych-113011-143823 View ArticlePubMedGoogle Scholar
- Bos, N., Groeneveld, C., van Bruggen, J., & Brand-Gruwel, S. (2016). The use of recorded lectures in education and the impact of lecture attendance and exam performance. British Journal of Educational Technology, 47, 906–917.View ArticleGoogle Scholar
- Butler, A. C., & Roediger, H. L. (2008). Feedback enhances the positive effects and reduces the negative effects of multiple-choice testing. Memory & Cognition, 36, 604–616. doi:10.3758/MC.36.3.604.View ArticleGoogle Scholar
- Cepeda, N. J., Pashler, H., Vul, E., Wixted, J. T., & Rohrer, D. (2006). Distributed practice in verbal recall tasks: a review and quantitative synthesis. Psychological Bulletin, 132, 354–380.View ArticlePubMedGoogle Scholar
- Chan, J. C. K., McDermott, K. B., & Roediger, H. L. (2006). Retrieval-induced facilitation: initially nontested material can benefit from prior testing of related material. Journal of Experimental Psychology: General, 135, 553–571.View ArticleGoogle Scholar
- Credé, M., Roch, S. G., & Kieszczynska, U. M. (2010). Class attendance in college: A meta-analytic review of the relationship of class attendance with grades and student characteristics. Review of Educational Research, 80, 272–295. doi:10.3102/0034654310362998 View ArticleGoogle Scholar
- Feld, G. B., & Diekelmann, S. (2015). Sleep smart – optimizing sleep for declarative learning and memory. Frontiers in Psychology, 6, 622. doi:10.3389/fpsyg.2015.00622 View ArticlePubMedPubMed CentralGoogle Scholar
- Grabe, M., & Christopherson, K. (2008). Optional student use of online lecture resources: resource preferences, performance and lecture attendance. Journal of Computer Assisted Learning, 24, 1–10.View ArticleGoogle Scholar
- Hartwig, M. K., & Dunlosky, J. (2012). Study strategies of college students: are self-testing and scheduling related to achievement? Psychonomic Bulletin and Review, 19, 126–134.View ArticlePubMedGoogle Scholar
- Karpicke, J. D., Butler, A. D., & Roediger, H. L. (2009). Metacognitive strategies in student learning: do student practice retrieval when they study on their own? Memory, 17, 471–479.View ArticlePubMedGoogle Scholar
- Kornell, N., Hays, M. J., & Bjork, R. A. (2009). Unsuccessful retrieval attempts enhance subsequent learning. Journal of Experimental Psychology: Learning, Memory & Cognition, 35, 989–998. doi:10.1037/a0015729 Google Scholar
- Lyle, K. B., & Crawford, N. A. (2011). Retrieving essential material at the end of lectures improves performance on statistics exams. Teaching of Psychology, 38(2), 94–97.View ArticleGoogle Scholar
- McDaniel, M. A., Agarwal, P. K., Huelser, B. J., McDermott, K. B., & Roediger, H. L. (2011). Test-enhanced learning in a middle school science classroom: the effects of quiz frequency and placement. Journal of Educational Psychology, 103, 399–414.View ArticleGoogle Scholar
- Pastötter, B., & Bäuml, K.-H. (2014). Retrieval practice enhances new learning: the forward effect of testing. Frontiers in Psychology, 5, 286. doi:10.3389/fpsyg.2014.00286 PubMedPubMed CentralGoogle Scholar
- Pennebaker, J. W., Gosling, S. D., & Ferrell, J. D. (2013). Daily online testing in large classes: boosting college performance while reducing achievement gaps. Plos One, 8(11), e79774.View ArticlePubMedPubMed CentralGoogle Scholar
- Putnam, A. L., Sunkhasattee, V. W., & Roediger, H. L. (2016). Optimizing learning in college: tips from cognitive psychology. Perspectives on Psychological Science, 11, 652–660. doi:10.1177/1745691616645770 View ArticlePubMedGoogle Scholar
- Richland, L. E., Kornell, N., & Kao, L. S. (2009). The pretesting effect: do unsuccessful retrieval attempts enhance learning? Journal of Experimental Psychology: Applied, 15(3), 243–257. doi:10.1037/a0016496 PubMedGoogle Scholar
- Roediger, H. L., Agarwal, P. K., McDaniel, M. A., & McDermott, K. B. (2011). Test-enhanced learning in the classroom: long-term improvements from quizzing. Journal of Experimental Psychology: Applied, 17(4), 382–395. doi:10.1037/a0026252 PubMedGoogle Scholar
- Roediger, H. L., & Karpicke, J. D. (2006). Test-enhanced learning. Psychological Science, 17, 249–255. doi:10.1111/j.1467-9280.2006.01693.x View ArticlePubMedGoogle Scholar
- Rohrer, D., Taylor, K., & Sholar, B. (2010). Tests enhance the transfer of learning. Journal of Experimental Psychology: Learning, Memory & Cognition, 365, 233–239. doi:10.1037/a0017678 Google Scholar
- Traphagan, T., Kucsera, J., & Kishi, K. (2010). Impact of class lecture webcasting on attendance and learning. Educational Technology Research and Development, 58, 10–37.View ArticleGoogle Scholar