- Original article
- Open Access
Optimizing learning in undergraduate psychology students: the impact of advance quizzing, review, and classroom attendance
Cognitive Research: Principles and Implications volume 2, Article number: 39 (2017)
This study investigates whether introducing simple cognitive interventions that are known to enhance learning in laboratory studies can be transferred to classroom settings. In an introductory psychology class, students were provided with a brief advance quiz on the topic covered in each particular lecture. In case they did not attend class, they still had the opportunity to fill it out at home. The interventions were offered on a voluntary basis and the students were given the opportunity to obtain an extra point towards the final exam by reviewing the key points of each lecture on a regular basis. The results indicate that both advance quizzing and review enhanced the grades on the final examination. Although it is likely that individual differences also contribute to learning outcomes, these effects were specific for the particular course and independent from class attendance. The results indicate that transfer of simple cognitive interventions from laboratory to classroom settings is feasible. They also indicate that these interventions can be effectively implemented into the lectures to boost student learning.
Undergraduate university students have only a fragmentary knowledge about effective learning strategies. In contrast, cognitive psychologists have a very fine-grained understanding about strategies and activities that benefit learning in laboratory studies. One of the most important insights is the effect of retrieval as an effective learning strategy. Retrieval benefits learning even for materials to be studied but not yet studied (i.e., advance quizzing). Retrieval induced as a review of key points between lectures also benefits learning. Transferring these strategies and activities into the classroom is an important mission for contemporary cognitive psychology. Although the effects of quizzing and review are small, they are consistent and specific to the learning materials.
Undergraduate university students have only a fragmentary knowledge about effective learning strategies (Bjork, Dunlosky, & Kornell, 2013; Karpicke, Butler, & Roediger, 2009; Putnam, Sunkhasattee, & Roediger, 2016). In contrast, cognitive psychologists have a fine-grained understanding about strategies and activities that benefit learning in laboratory studies. For example, there is convincing evidence that testing is a very effective learning strategy (Roediger & Karpicke, 2006). In fact, even unsuccessful retrieval attempts can enhance subsequent learning (Kornell, Hays, & Bjork, 2009; Richland, Kornell, & Kao, 2009). In order to test whether these findings transfer to undergraduate university student learning in a classroom setting, the present study used a brief quiz at the beginning of each lecture on the topic to be covered in that particular lecture. Beside this intervention, which focused on retrieval before the lecture, a second intervention was administered that focused on retrieval after the lecture. Specifically, the students were given the opportunity to obtain an extra point towards the final exam by reviewing the key points of each lecture on a regular basis.
There is already some evidence that quizzing as a learning tool can be successfully used in the classroom. For example, regularly answering a set of questions that required the retrieval of information from the same day’s class significantly improved subsequent exam performance in a statistics course (Lyle & Crawford, 2011). In an introductory psychology class, Pennebaker, Gosling, and Ferrell (2013) presented a brief computerized quiz using an online system to test for the content of the previous lesson. Their results showed better performance and moreover, a reduction in the gap between students from different social classes. Using an experimental approach in the classroom, Roediger, Agarwal, McDaniel, and McDermott (2011) showed that quizzing had a beneficial effect for middle school students in a social studies course. Similarly, McDaniel, Agarwal, Huelser, McDermott, and Roediger (2011) demonstrated that quizzing improved learning in an eighth-grade science class. In one experiment, they specifically manipulated the placement of quizzing, with some students being tested before the respective lecture was introduced in class. The results showed that, overall, review quizzing produced the greatest increase in final exam performance, advance quizzing also seemed to provide a benefit compared to not having a quiz at all. In the present study, the impact of advance quizzing is addressed in university students, notably without a strict requirement for students to complete every activity.
Although it may seem counterintuitive to answer questions about topics that have not been studied yet, answering questions beforehand activates related knowledge about the topic and makes it easier to connect new information to what is already known. Moreover, it can raise the interest in the topic and thus enhance attention during the lecture (Richland et al., 2009). Thus, it can have both direct and indirect effects. In the present study, four questions were prepared for each lecture on the university online learning platform. This allowed giving individual feedback at the end of the quiz. As classroom attendance was facultative, a podcast of the lecture that contained the link to the advance quiz was made available on the learning platform. By considering the access time to the quiz, it was possible to identify students who attended the class, students who used the online course materials, and those who did neither. Therefore, the study can also contribute to the question whether attending class and/or the availability of recorded lectures is associated with performance on the final exam (cf. Credé, Roch, & Kieszczynska, 2010; Bos, Groeneveld, van Bruggen, & Brand-Gruwel, 2016; Traphagan, Kucsera, & Kishi, 2010).
The study further investigated whether reviewing the material after the lecture would benefit learning. Specifically, students were motivated to review each lecture and to submit the three key points to the learning platform before the next lecture. For the regular submission, they received a bonus point towards the final exam. The idea behind this intervention was to provide the students with retrieval practice and thus to facilitate learning (cf. Roediger & Karpicke, 2006). Moreover, reviewing after a delay can also be considered as spaced practice, which is also known to facilitate learning (Cepeda, Pashler, Vul, Wixted, & Rohrer, 2006). As sleep enhances memory consolidation, the time window for submission did not open until the day after the lecture. Moreover, in line with the reconsolidation hypothesis, which proposes that when a memory is reactivated, the trace may return to a labile state after which time it is reconsolidated or restabilized, the idea was that the requirement for reconsolidation would further strengthen learning (Feld & Diekelmann, 2015).
The interventions were embedded in an introductory psychology course (i.e., “Introduction to Memory”) that students had to attend as part of the curriculum. Due to a change in the curriculum, a larger number of students – from the first and from the second year – were enrolled in the course (i.e., two study years). For the first-year students, the curriculum change involved a new testing format in which the exam was combined with another cognitive psychology introduction course (“Introduction to Perception”) which had taken place one semester earlier. For the benefit of the present study, the performance achieved in this other part of the exam could be used to test the specificity of the intervention effects.
Settings and participants
Participants were undergraduate psychology students from the University of Bern who took the course “Introduction to Memory” as part of their curriculum. The study was conducted in agreement with the ethical guidelines of the Human Science Faculty of the University of Bern. The original number of students enrolled in the course was 673; a total of 300 first-year students and 210 second-year students took the final exam and were included in the analyses. Due to the change in the curriculum two student cohorts (i.e., first-year and second-year students) were taught together. Related to this change, the exam of the first-year students was combined with another course (“Introduction to Perception”), which had taken place one semester earlier. The interval between the last class of that course and the final exam was 26 weeks. For first-year students, performance achieved in this exam was used to test the specificity of the advance quizzing and review interventions. Consequently, the data on first- and second-year students were analysed separately.
Measures and materials
The “Introduction to Memory” course consisted of thirteen lectures distributed across 13 weeks with one lecture per week, and was, to a large part, based on the book by Baddeley, Eysenck, and Anderson (2015). Each lecture was recorded and made available as a podcast on the online learning platform of the university (ilias.unibe.ch). Moreover, the slides for each lecture were uploaded before each lecture, in accordance with the recommendation by Putnam et al. (2016).
For each lecture, four multiple choice questions were prepared. These were accessible via the online learning platform. The questions covered some of the content of the actual lecture and provided immediate feedback about the correct solution (cf. Richland et al., 2009; Butler & Roediger, 2008). These were arranged in two different access windows. The first one was open during the lecture and was used as a proxy to assess classroom attendance. In the first part of each lecture, students were provided with about 5 minutes to answer these questions in class. The second window opened after the lecture and thus the quiz was also available for those students who did not attend the lecture, but studied the particular topic using the online materials (e.g., slides, podcasts etc.). For the purpose of this study, the number of completed quizzes was used. Thus, it was possible to have a maximum score of 13 both for the lecture time window (which was used as a proxy for classroom attendance) and for the time window that opened only after the lecture (which was used as a proxy for podcast use).
In order to motivate the students to review the content of a lecture before the next one, students were given the opportunity to submit three key points via the online learning platform. In order to benefit from overnight consolidation (cf. Feld & Diekelmann, 2015), the opportunity to submit the review opened the day after the lecture and lasted until the day before the next lecture. Thus, it was possible to have a maximum of 13 completed reviews. An extra point towards the final exam was given when at least 12 reviews were submitted via the online platform. This additional point was considered to calculate the final grade. However, for the purpose of this study, it was not added to the points in the final exam.
The final exam took place 3 weeks after the last lecture. It consisted of 22 multiple choice questions, which covered the content of all the lectures (i.e., 1–2 questions per topic). For the first-year students the exam was combined with the exam on the course “Introduction to Perception”, which was prepared accordingly. For each exam a maximum of 22 points was possible. As noted above, the extra point that could be gained by submitting the reviews on a regular basis was not considered to calculate the points for the final exam for the purpose of this study.
In the results section, first the descriptive statistics for each variable are presented. Next, a group comparison is provided to test for differences between first-year and second-year students. The main focus is on the relationship between advance quizzing, review and performance in the final exam, and correlations are provided separately for first-year and second-year students. To test the predictive value of these interventions, separate regression analyses were conducted. For each of these analyses, diagnostic statistics were run to check whether the assumptions were met. The results of the diagnostics are summarized here. Analyses of standard residuals showed that the data contained no outliers (standard residuals < 3). To test whether the data met the assumption of collinearity, VIF scores were calculated, with values between 1.1 and 1.5, indicating that multicollinearity was not a concern. The data also met the assumption of independent errors (with Durbin-Watson values = 2 +/− 0.5). Histograms of standardized residuals and the plots of standardized residuals showed that the errors were approximately normally distributed. The scatterplot of standardized residuals showed that the data also met the assumptions of homogeneity of variance and linearity. The data also met the assumption of non-zero variances. In a set of follow-up analyses, the exclusivity of the contribution of advance quizzing was assessed using a stricter criterion. Using hierarchical regression analyses, the extra contribution of advance quizzing in the classroom was tested after controlling for the influence of quizzing after the lecture and review.
Table 1 shows descriptive statistics for each variable. Although the main focus was on the relationship between advance quizzing, review and performance in the final exam, a first set of analyses tested for differences between the two student groups. As noted in Table 1, second-year students seemed to attend the classroom less often than first-year students. However, they used the advance quizzing opportunity that was available after the lecture more often. These results are also illustrated in Fig. 1, which depicts the number of advance quizzes that were answered in the lecture (i.e., classroom attendance) and after the lecture (i.e., podcast use). There was no group difference in the number of reviews that were submitted via the learning platform. These results are illustrated in Fig. 2. In the final exam, second-year students performed better than first-year students (Table 1). This may be due to the fact that first-year students had to complete more concurrent exams than second-year students.
In order to examine the relationship between advance quizzing, review and performance in the final exam, correlation was separately tested for first-year and second-year students, respectively. The correlation coefficients are presented in Table 2. Correlation between using advance quizzing during the class vs. using it after the class was weak in both groups. However, in both groups there were moderate correlation between advance quizzing and the scores in the final exam for the “Introduction to Memory” course, in which these interventions were introduced. Notably, correlation was weak for the “Introduction to Perception” course, in which these interventions were not implemented.
To test the predictive value of these interventions, three separate regression analyses were conducted. The first analysis focused on the first-year students. The second analysis focused on the second-year students to establish the generality of the results. The third analysis focused on the specificity of the influences of advance quizzing and review by testing whether the interventions would also predict the performance of the first-year students in the “Introduction to Perception” course.
The first analysis which focused on the first-year students included advance quizzing in the classroom, and after class (supposedly when using the podcast), and review as predictors and the “Introduction to Memory” exam as the criterion. The predictor variables explained a significant amount of variance of the criterion variable, F(3, 296) = 19.84, p < .001, R 2 = .167. The regression coefficients are presented in Table 3. They show that advance quizzing in the classroom, quizzing after class, and review all contributed significantly to predict the exam score.
The second analysis focused on the second-year students and included the same variables. Again, the predictors explained a significant amount of variance in the criterion variable, F(3, 206) = 13.41, p < .001, R 2 = .163. The regression coefficients are presented in Table 4. Again, advance quizzing in the classroom, quizzing after class, and review all contributed significantly to predict the exam score, thus replicating the findings from the first-year students.
To test for the specificity of the interventions, the “Introduction to Perception” exam was used as the dependent variable in the third analysis. If the results of the interventions found in the first two regression analyses are not simply artefacts, then the amount of variance explained should be much smaller. The results showed that the predictor variables still explained a significant amount of variance in the criterion variable, F(3, 296) = 5.23, p < .01, R 2 = .050. However, the amount of explained variance was low (i.e., only 5% compared to 16–17% for the memory exam). As can be seen in Table 5, only the regression coefficient of advance quizzing in the classroom had a significant effect. Neither advance quizzing after class nor review contributed significantly to predict the exam score. This demonstrates the specificity of the effects for the memory course.
If one considers classroom attendance as a stable trait of a person (e.g., conscientiousness) and if only classroom attendance would predict exam sores, one would expect the same predictive power of advance quizzing in the classroom variable for both the memory and the perception exam. However, a comparison of the standardized beta coefficients shows that the predictive power of this variable is higher for the memory exam, that is, .177 (Table 3) and .180 (Table 4) for first-year and second-year students, respectively, compared to .134 for the perception exam (Table 5). This indicates that not only classroom attendance contributes to this relationship, but also that advance quizzing in the classroom provides additional explanatory power.
The previous analyses evaluated the impact of the interventions assuming that the advance quizzes were always used before learning. However, it is possible that the students who did not attend class may have completed the quizzes after they studied. If so, then the quizzes would not be advance quizzes for those students. It is also possible that students accessed the quizzes during class time and used them as study guides during the lecture rather than using them at the beginning of the lecture as instructed. Again, if so, the quizzes would not be advance quizzes for those students. In order to control for these possibilities further analyses were conducted. In these analyses, the focus was on those students who completed the advance quizzes during class, and specifically, during the first part of the lecture as instructed. The latter was possible due to the availability of the exact time stamp of the access to the advance quiz. To allow for a proper measurement of the impact of advance quizzing at the beginning of the lecture, instances in which students also accessed the advance quiz after the lecture were excluded. As no time stamp information was available for the access to the podcast it was not possible to apply the same restrictions to those students who did not attend class. With these restrictions, the mean number of uses of advance quizzing during the lecture was 3.91 (SD = 3.65) for first-year students and 2.79 (SD = 3.60) for second-year students, which was statistically significant, t(508) = 3.44, p < .01.
To test whether advance quizzing during the lecture had an exclusive predictive value for the final exam, hierarchical regression analyses were conducted that included quizzing after the lecture and review in the first step and advance quizzing at the beginning of the lecture in the second step, separately for first-year students and second-year students.
The results of these analyses are presented in Tables 6 and 7, for first-year and second-year students, respectively. Consistently, the results of these follow-up analyses showed that advance quizzing before learning contributed to performance in the final exam. Specifically, the inclusion of advance quizzing before learning led to a 16% increase in the amount of explained variance (.022/.14) for first-year students and a 26% increase of the amount of explained variance for second-year students (.036/.136), respectively. These results demonstrate that advance quizzing had an additional and independent effect on memory exam performance.
The purpose of the present study was to investigate whether retrieval practice before learning (i.e., advance quizzing) and retrieval practice after learning (i.e., review between lectures) would affect final exam grades in an introductory psychology class. The results indicated that both variables enhanced final grades. Although the effects were small, they were consistent and replicated in two separate student groups (i.e., first-year and second-year students). Moreover, the effect was specific as it was not present in another cognitive psychology introduction course attended by the same first-year students.
These results are in line with other studies that showed beneficial effects of quizzing, retrieval practice, and spaced retrieval (e.g., McDaniel et al., 2011; Richland et al., 2009; Roediger & Karpicke, 2006). It is likely that these interventions have both direct and indirect effects on final exam performance. The specific processing of study materials can lead to greater familiarity and reduce anxiety in particular for complex materials. Quizzing and retrieval can lead to more elaborate associations and retrieval cues, can enhance consolidation, and lead to more elaborate memory traces through reconsolidation. Repeated processing of the materials can also lead to retrieval-induced facilitation for other materials and enhanced transfer of learning (Chan, McDermott, & Roediger, 2006; Rohrer, Taylor, & Sholar, 2010). Advance quizzing may specifically help in building up new memory representations. It may function as an advance organizer that guides attention through the lectures and thus enhances attentiveness. Importantly, due to the voluntary nature of the interventions in this study, self-initiated processes may be particularly boosted. The interventions can also help to optimize students’ learning strategies, e.g., helping students identify what they do not understand, which in turn can trigger further studying, and guide them to schedule their study time (Bjork et al., 2013; Hartwig & Dunlosky, 2012; Karpicke et al., 2009; Putnam et al., 2016).
Overall, the results of the current study indicate that even rather simple interventions based on the insights from basic research in laboratory cognitive psychology can have a significant impact on student learning in the classroom. The results confirm theoretical considerations from laboratory studies on the effects of testing, pretesting, and the forward effect of testing (e.g., Roediger & Karpicke, 2006, Kornell et al., 2009; Richland et al., 2009; Pastötter & Bäuml, 2014). In contrast to previous work, in the present study, the interventions were conceptualized on a voluntary basis, that is, students were not strictly required to participate in the interventions. Thus, the quality of the submitted responses, in particular for the review intervention, may have varied substantially. Nevertheless, participating on a regular basis still improved overall course performance.
The results also shed light on whether or not classroom attendance is necessary for successful performance. They indicate that classroom presence is not a mandatory precondition for success in the final exam. Rather it seems that using the provided materials on the online platform such as the podcast and slides can compensate for the lack of “live” experience. These results are in line with other studies that have found no detrimental effects of new technologies compared to traditional face-to-face lectures (e.g., Bos et al., 2016; Grabe & Christopherson, 2008). They are also encouraging for the growing field of distance education where classroom attendance is not possible at all.
However, a limitation of the present study is the measurement of classroom attendance. It is important to note that advance quizzing in the classroom was used as a proxy for classroom attendance. As some students did not bring their computers/smartphones to class, this measure is not comprehensive. Similarly, using advance quizzing after the class as a proxy for podcast use is also imprecise. In fact, the specific use of podcasts was not assessed at all. Nevertheless, the results indicate that advance quizzing during class and advance quizzing after class both contributed to predicting the final exam performance.
On a related note, it is possible that students who did not attend class may have completed the quizzes after they studied. If so, then the quizzes would not be advance quizzes for those students. Similarly, it is possible that students accessed the quizzes during class time and used them as study guides during the lecture rather than using them before learning. If so, the quizzes would not be advance quizzes for those students. In order to control for these possibilities further analyses were conducted to test for the independent contribution of advance quizzing. In these analyses, the focus was on those students who completed the advance quizzes during class, and specifically, during the first part of the lecture as instructed. In addition, instances in which students also accessed the advance quiz after the lecture were excluded. The results of these analyses showed that advance quizzing before learning contributed exclusively to performance in the final exam over and above the other two variables.
A caveat may also apply to the review intervention. As the specific review submissions were not controlled for quality or accuracy, they also represent a rather vague variable. It is very possible that the predictive value of review would have been even stronger if such a quality control had been implemented. This may be an avenue for future research. Despite these shortcomings, the presence of a consistent enhancing effect is striking.
Finally, due to the correlational approach, it is likely that extraneous variables such as individual differences also influenced the outcome. Specifically, more conscientious students may study more, may be more likely to participate in the offer of learning aids, and may attend class more often. Although the correlation between advance quizzing and final exam performance was much stronger for the “Introduction to Memory” class than for the “Introduction to Perception” class, the latter correlation was still significant. This supports the hypothesis that individual differences also contribute to the learning outcomes.
To conclude, this study demonstrates that simple cognitive interventions transfer from laboratory to classroom settings. Retrieval benefits learning even for materials to be studied but not yet studied (i.e., advance quizzing). Retrieval induced as a review of key points between lectures also benefits learning. Transferring these strategies and activities into the classroom is an important mission for contemporary cognitive psychology. The results indicate that some effective interventions can be implemented rather easily into the lectures to boost student learning.
Baddeley, A. D., Eysenck, M. W., & Anderson, M. C. (2015). Memory. NY: Psychology Press.
Bjork, R. A., Dunlosky, J., & Kornell, N. (2013). Self-regulated learning: beliefs, techniques, and illusions. Annual Review of Psychology, 64, 417–444. doi:10.1146/annurevpsych-113011-143823
Bos, N., Groeneveld, C., van Bruggen, J., & Brand-Gruwel, S. (2016). The use of recorded lectures in education and the impact of lecture attendance and exam performance. British Journal of Educational Technology, 47, 906–917.
Butler, A. C., & Roediger, H. L. (2008). Feedback enhances the positive effects and reduces the negative effects of multiple-choice testing. Memory & Cognition, 36, 604–616. doi:10.3758/MC.36.3.604.
Cepeda, N. J., Pashler, H., Vul, E., Wixted, J. T., & Rohrer, D. (2006). Distributed practice in verbal recall tasks: a review and quantitative synthesis. Psychological Bulletin, 132, 354–380.
Chan, J. C. K., McDermott, K. B., & Roediger, H. L. (2006). Retrieval-induced facilitation: initially nontested material can benefit from prior testing of related material. Journal of Experimental Psychology: General, 135, 553–571.
Credé, M., Roch, S. G., & Kieszczynska, U. M. (2010). Class attendance in college: A meta-analytic review of the relationship of class attendance with grades and student characteristics. Review of Educational Research, 80, 272–295. doi:10.3102/0034654310362998
Feld, G. B., & Diekelmann, S. (2015). Sleep smart – optimizing sleep for declarative learning and memory. Frontiers in Psychology, 6, 622. doi:10.3389/fpsyg.2015.00622
Grabe, M., & Christopherson, K. (2008). Optional student use of online lecture resources: resource preferences, performance and lecture attendance. Journal of Computer Assisted Learning, 24, 1–10.
Hartwig, M. K., & Dunlosky, J. (2012). Study strategies of college students: are self-testing and scheduling related to achievement? Psychonomic Bulletin and Review, 19, 126–134.
Karpicke, J. D., Butler, A. D., & Roediger, H. L. (2009). Metacognitive strategies in student learning: do student practice retrieval when they study on their own? Memory, 17, 471–479.
Kornell, N., Hays, M. J., & Bjork, R. A. (2009). Unsuccessful retrieval attempts enhance subsequent learning. Journal of Experimental Psychology: Learning, Memory & Cognition, 35, 989–998. doi:10.1037/a0015729
Lyle, K. B., & Crawford, N. A. (2011). Retrieving essential material at the end of lectures improves performance on statistics exams. Teaching of Psychology, 38(2), 94–97.
McDaniel, M. A., Agarwal, P. K., Huelser, B. J., McDermott, K. B., & Roediger, H. L. (2011). Test-enhanced learning in a middle school science classroom: the effects of quiz frequency and placement. Journal of Educational Psychology, 103, 399–414.
Pastötter, B., & Bäuml, K.-H. (2014). Retrieval practice enhances new learning: the forward effect of testing. Frontiers in Psychology, 5, 286. doi:10.3389/fpsyg.2014.00286
Pennebaker, J. W., Gosling, S. D., & Ferrell, J. D. (2013). Daily online testing in large classes: boosting college performance while reducing achievement gaps. Plos One, 8(11), e79774.
Putnam, A. L., Sunkhasattee, V. W., & Roediger, H. L. (2016). Optimizing learning in college: tips from cognitive psychology. Perspectives on Psychological Science, 11, 652–660. doi:10.1177/1745691616645770
Richland, L. E., Kornell, N., & Kao, L. S. (2009). The pretesting effect: do unsuccessful retrieval attempts enhance learning? Journal of Experimental Psychology: Applied, 15(3), 243–257. doi:10.1037/a0016496
Roediger, H. L., Agarwal, P. K., McDaniel, M. A., & McDermott, K. B. (2011). Test-enhanced learning in the classroom: long-term improvements from quizzing. Journal of Experimental Psychology: Applied, 17(4), 382–395. doi:10.1037/a0026252
Roediger, H. L., & Karpicke, J. D. (2006). Test-enhanced learning. Psychological Science, 17, 249–255. doi:10.1111/j.1467-9280.2006.01693.x
Rohrer, D., Taylor, K., & Sholar, B. (2010). Tests enhance the transfer of learning. Journal of Experimental Psychology: Learning, Memory & Cognition, 365, 233–239. doi:10.1037/a0017678
Traphagan, T., Kucsera, J., & Kishi, K. (2010). Impact of class lecture webcasting on attendance and learning. Educational Technology Research and Development, 58, 10–37.
I would like to thank Stefan Walter for technical support and Janek Lobmaier for providing access to the perception exam scores.
This project did not have a funding source.
Availability of data and materials
The data and materials are made available in Additional file 1.
Ethics approval and consent to participate
The study was conducted in agreement with the ethical guidelines of the Human Science Faculty of the University of Bern.
Consent for publication
The author declares that he has no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Meier, B. Optimizing learning in undergraduate psychology students: the impact of advance quizzing, review, and classroom attendance. Cogn. Research 2, 39 (2017). https://doi.org/10.1186/s41235-017-0075-2
- Testing effects
- Pretesting effects