Assessing Long-Term Effects of Inquiry-Based Learning: A Case Study from College Mathematics

Innovative Higher Education, Jun 2014

As student-centered approaches to teaching and learning are more widely applied, researchers must assess the outcomes of these interventions across a range of courses and institutions. As an example of such assessment, this study examined the impact of inquiry-based learning (IBL) in college mathematics on undergraduates’ subsequent grades and course selection at two institutions. Insight is gained upon disaggregating results by course type (IBL vs. non-IBL), by gender, and by prior mathematics achievement level. In particular, the impact of IBL on previously low-achieving students’ grades is sizable and persistent. The authors offer some methodological advice to guide future such studies.

A PDF file should load here. If you do not see its contents the file may be temporarily unavailable at the journal website or you do not have a PDF plug-in installed and enabled in your browser.

Alternatively, you can download the file locally and open with any standalone PDF reader:

https://link.springer.com/content/pdf/10.1007%2Fs10755-013-9269-9.pdf

Assessing Long-Term Effects of Inquiry-Based Learning: A Case Study from College Mathematics

Marina Kogan Sandra L. Laursen 0 ) Ethnography & Evaluation Research, University of Colorado Boulder , 580 UCB, Boulder, CO 80309-0580, USA As student-centered approaches to teaching and learning are more widely applied, researchers must assess the outcomes of these interventions across a range of courses and institutions. As an example of such assessment, this study examined the impact of inquiry-based learning (IBL) in college mathematics on undergraduates' subsequent grades and course selection at two institutions. Insight is gained upon disaggregating results by course type (IBL vs. non-IBL), by gender, and by prior mathematics achievement level. In particular, the impact of IBL on previously low-achieving students' grades is sizable and persistent. The authors offer some methodological advice to guide future such studies. - Marina Kogan has bachelors degrees in sociology/anthropology and computer science from City University of New York and an M.A. in sociology from the University of Illinois Urbana-Champaign. She conducted this work as a researcher at Ethnography & Evaluation Research at the University of Colorado Boulder and is now pursuing a doctorate in computer science from the University of Colorado Boulder, with research interests in the application of computer modeling and data mining methods to social problems. Sandra L. Laursen has undergraduate degrees in chemistry and French from Grinnell College and a Ph.D. in chemistry from the University of California-Berkeley. She is the Co-director of Ethnography & Evaluation Research at the University of Colorado Boulder and leads evaluation and research studies of higher education and career paths in science, engineering, and mathematics. She can be reached at . To date, the most persuasive studies of active learning have examined student outcomes within a single course at one or several institutions (e.g., Deslauriers et al., 2011; Kwon, Rasmussen & Allen, 2005; and studies analyzed by Froyd, 2008, and Ruiz-Primo et al., 2011). However, as active learning approaches are applied more broadly, evaluating their outcomes presents new methodological challenges. Measures to assess effectiveness must be general enough to apply across different classrooms and institutions. A common testthe most direct method of evaluating classroom learningmay not be available or applicable. Students course grades and course-taking patternstheir choices to pursue (or not) subsequent courses in a disciplineoffer broad and arguably objective measures for evaluating the effects of an educational intervention. While grading standards differ across instructors, courses, and campuses, grades have a fairly stable social meaning (Pattison, Grodsky & Muller, 2013). As part of students academic transcripts, grades become lasting records of achievement. Like grades, course-taking patterns apply to varied academic contexts; they may reflect students sustained or lost interest in a discipline following an initial experience. Several recent studies have used various grade- and course-taking measures to evaluate the success of an educational intervention, including final grades and pass/fail rates (e.g., Dubetz et al., 2008; Tai, Sadler & Mintzes, 2006; Tien, Roth & Kampmeier, 2002), the next grade in a course sequence (e.g., Farrell, Moog, & Spencer, 1999; Gafney & Varma-Nelson, 2008), grades in multiple subsequent courses (e.g., De Paola, 2009; Weinberg, Hashimoto, & Fleisher, 2009), and enrollment in higher level electives (Carrell & West, 2010). Mostrom and Blumberg (2012) suggested that student-centered courses are subject to accusations of grade inflation because the course has lost content or rigor or because different assessment methods enable students to do better. They also argued that grade improvement may in fact measure real improvement in learning. Measures that focus on subsequent courses avoid this issue because students who did and did not experience the intervention all take the same later courses. Moreover, such measures can detect valued and lasting impact on students learning, academic success, or academic choices (Derting & Ebert-May, 2010). This study examined undergraduates grades and course-taking following an inquirybased learning (IBL) experience in college mathematics. In the context of mathematics, IBL approaches engage students in exploring mathematical problems, proposing and testing conjectures, developing proofs or solutions, and explaining their ideas. As students learn new concepts through argumentation, they also come to see mathematics as a creative human endeavor to which they can contribute (Rasmussen & Kwon, 2007). Consistent with current socio-constructivist views of learning, IBL methods emphasize individual knowledge construction supported by peer social interactions (Ambrose et al., 2010; Cobb, Yackel & McCain, 2000; Davis, Maher & Noddings, 1990). In this article we report our analysis of student academic records for patterns in grades and course-taking among students who had earlier taken an IBL mathematics course or comparative, non-IBL course taught with other methods. We focus on results for two groups often under-served by traditionally taught college mathematics courses, women and low-achieving students. The academic records study was one element of a large, mixed-methods study of IBL mathematics as implemented at four universities hosting IBL Math Centers (Laursen, Hassi & Hough, 2013; Laursen, Hassi, Kogan, Hunter & Weston, 2011; Laursen, Hassi, Kogan & Weston, 2013). Observation, survey, interview, and test data were gathered from over 100 sections of 40 courses aimed at varied levels and audiences. First we describe results of classroom observations which establish that IBL was a student-centered, educational intervention. We then outline the methods used to study subsequent grades and course-taking for students who had completed an IBL course or its non-IBL counterpart (for details see Laursen et al., 2011). Setting and Courses Each of the four institutions selected and developed its IBL courses independently and labeled them as IBL or non-IBL based on instructor participation in their grant-funded IBL Center. The courses were well established, having been taught several times prior to our data collection in 2009. To check these labels and to establish whether observed differences in student outcomes were meaningful, we carried out over 300 hours of classroom observation of 42 course sections, having received human subjects approval from our Universitys Institutional Review Board and that of each study site where required. The results showed that, despite variation among courses and instructors, several key characteristics differentiated the IBL courses from the non-IBL courses. On average, about 60% of class time in IBL courses was spent on student-centered activities such as small group work, student presentation of problems at the board, or whole-class discussion, while in non-IBL courses over 85% of class time consisted of the instructor talking. In IBL courses, students more often took on leadership roles and asked more questions. Trained observers rated IBL courses higher for creating a supportive classroom atmosphere, eliciting student intellectual input, and providing feedback to students on their work. Overall, the data clearly show that students who took IBL sections experienced a different instructional approach than those in lecture-based, non-IBL sections (Laursen et al., 2011, 2013b). The academic records study focused on two Centers that offered high-enrollment courses taught in both IBL and non-IBL formats. We chose three target courses based on the following: Placement early enough in a typical course sequence to allow for variation in subsequent course choices and grades, Target sections taught in prior years early enough that subsequent course-taking was near complete at the time of data collection in 2009, and Adequate numbers of students enrolled in both IBL and non-IBL sections. The two courses at Center L were a middle-level introduction to proof course (designated L1) and an advanced proof-based course (L2). L1 aimed to help students shift from the problem-solving of calculus to the rigorous proof-based approach of advanced courses. It met degree requirements for mathematics, some science and engineering fields, and secondary math teaching. Course L2 was not required but counted toward the math major. Both L1 and L2 were taught in sections of 20-30 students. Course sections were not institutionally labeled as IBL or non-IBL; self-selection occurred but was not extensive. Therefore we used statistical methods to control for differences among entering students that might affect their later academic outcomes. The third course, G1, was the first course in a three-term sequence including multivariable calculus, linear algebra, and differential equations; all three were offered in IBL and non-IBL formats. Both institutional selection and student self-selection operated heavily in this course. Students were invited to join the IBL honors section based on past mathematics performance, thus populating these sections with high-achieving, self-motivated students. Non-IBL sections included students of all prior achievement levels taught in large lectures with recitations led by graduate teaching assistants. On average, IBL students had higher SAT scores and high school GPAs than non-IBL students, took G1 earlier (often in their first college term), and pursued mathematics majors in higher numbers. To compare these groups fairly, we constructed a matched sample, which is detailed below. We considered several measures by which to assess academic outcomes, beginning with anonymized raw data from standard institutional records. DFW rates, the proportion of students who fail (earn D or F grades) or withdraw (W) from a class (Dubetz et al., 2008), were not useful since they are low in honors and upper-level courses. Instructors argued that grades and exam scores could not be compared across IBL and non-IBL sections, given differences in emphasis and assessment. Instead, we developed standardized approaches to counting and averaging grades in courses taken after the target course. Because students who took IBL or non-IBL sections of the target later co-enrolled in other courses, their later grades can be compared directly to each other, albeit not across courses or institutions. Here we outline the standardized variables. Variables describing course counts focus on new courses completed. Repeated courses were counted if the initial attempt ended in withdrawal or a failing grade. Because many students had not graduated by the time of data collection, we analyzed only courses taken within two years after the target course. Course-counting variables included the following: Number of prior math coursesbefore the target course, a control for math background; Number of subsequent math coursesall courses taken after the target course; Number of subsequent elective courseselective courses taken after the target course, other than core courses required for the mathematics major; and Number of subsequent IBL coursesIBL-method courses taken after the target course. The number of subsequent required courses is largely determined by students progress in the major and not a useful measure of student choice. Major-switching in or out of mathematics was minimal for all groups in these courses. Variables to describe math grades focus on courses for which a letter grade was received. Differently from the count variables, all letter grades were included in calculating averages, including repeated courses. Grade variables included the following: Average prior gradein courses before the target course, a control for math achievement; Next term average grade; and Average grade in subsequent elective, required, and IBL courses. For the average grade variables, sample sizes differed, since not all students took all types of courses. For example, most non-IBL students took no IBL courses. Statistical control procedures further reduced sample sizes as not all student records contained the data used to control for incoming differences. Table 1 shows the maximum sample sizes for each course; actual sample sizes for each grade variable are included with the detailed results in Tables 2 and 3. To account for students prior achievement or ability entering the first-year course G1, we created an index combining students high school GPA with college admissions test scores. Concordance tables were used to convert ACT to SAT mathematics scores and ACT English and reading scores to SAT verbal scores (Dorans et al., 1997; Dorans, 1999, 2004). Because high school grades and admissions test scores are comparably important predictors of college success (Hoffman & Lowitzki, 2005; Noble, 1991), the index weighted high school GPA, math SAT score, and verbal SAT score approximately equally. The new index was divided into seven equal brackets. Major demographic sub-groups 23% soph 71% M 52% white 28% jr 29% F 21% Asian 39% sr 12% Hispanic 26% jr 65% M 51% white 52% sr 35% F 19% Asian 16% Hispanic 10% foreign 72% first 62% M 64% white 24% soph 38% F 18% Asian 10% Hispanic Table 1 Study Samples for Academic Records Analyses Method of control for incoming differences L1: Mid-level statistical L2: Upper-level statistical G1: Introductory sampling, 197 (of 962) 49 statistical Sampling and Analysis Table 1 summarizes the demographic information for samples from the three courses, totaling 3,212 students. Non-IBL samples were larger because IBL section offerings were limited. To develop a non-IBL sample comparable to the selective population in IBL sections of G1, we used the pre-college index plus demographic variables to match students. For each IBL student we selected two non-IBL students matched by index bracket, academic major (math, science, non-STEM, undeclared), academic status (freshman-senior), gender, and race/ethnicityin that order of priority. Overall, this process yielded highly similar IBL and non-IBL samples. SPSS (version 18) was used for statistical analyses. To compare means for IBL and nonIBL students we used primarily non-parametric tests (Mann-Whitney, Kruskal-Wallis, Chisquare), as most of the data were not normally distributed. We found some incoming differences between IBL and non-IBL student groups in the number of math courses and average prior math grade. For L1, IBL students had taken fewer prior math courses and earned higher average math grades prior to the target course. For L2, these differences were not significant. For G1, even after our close-match sampling, there was still a significant difference in the number of prior math courses, with IBL students taking fewer. Thus all reported results are based on applying the General Linear Model (GLM) procedure in SPSS to control for these incoming differences, using as covariates the number of math courses and average prior math grade, or for G1, the pre-college index. We report estimated marginal means, which are intended to offset the effect of the covariates as intervening variables. Effect sizes for the IBL intervention were computed from estimated marginal means and pooled standard deviations for all students and by gender. A different approach was required for effect sizes by prior achievement group. The GLM procedure adjusts the post-intervention student outcomes by controlling for prior math GPA. Because the achievement subgroups are based on prior math GPA, using these adjusted outcome measures to calculate effect sizes by achievement group would obfuscate precisely the group differences of interest. Instead, we used Morris (2008) Pretest-Posttest-Control group design. This method controls for preexisting differences even when treatment and control groups are nonequivalent, by allowing each individual to be used as his or own control, which typically increases the power and precision of statistical tests (p. 365). This design is only appropriate for the grade variables, as the number of math courses taken prior to the intervention does not have a pretest relationship to subsequent course counts; it is a measure of student preparedness instead. Sample size (overall & all course counts) Average grade in the next term N 351 Mean 2.816 standard deviation 0.918 Average grade in subsequent required courses N 336 Mean 2.626 standard deviation 0.917 Average grade in subsequent elective courses N 507 Mean 2.790 standard deviation 0.878 Average grade in subsequent IBL courses N 51 Mean, 2.523 standard deviation 1.000 Sig., IBL vs. non-IBL Effect size for IBL intervention Number of subsequent required courses Mean 0.473 standard deviation 0.632 Sig., IBL vs. non-IBL Sig., IBL vs. non-IBL, within gender Effect size for IBL intervention Sig., men vs. women ** Number of subsequent elective courses Mean 1.768 standard deviation 1.841 Number of subsequent IBL courses Mean 0.059 standard deviation 0.275 Sig., IBL vs. non-IBL Sig., IBL vs. non-IBL, within gender ** Effect size for IBL intervention Course L2, 2002-2008 Table 2 Estimated marginal means for average grade and courses taken subsequent to an IBL or non-IBL mathematics course, for all students and by gender, for three courses * 0.320 * Table 2 (continued) Average grade in subsequent required courses N 92 Mean 2.214 standard deviation 1.045 Average grade in subsequent elective courses N 289 Mean 2.642 standard deviation 0.918 Average grade in subsequent IBL courses N 15 Mean, 2.194 standard deviation 0.918 Number of subsequent required courses Mean 0.051 standard deviation 0.197 Number of subsequent elective courses Mean 1.399 standard deviation 1.594 Number of subsequent IBL courses Mean standard deviation Course G1, 2004-2006 Sample size (overall & all course counts) Average grade in the next term N 47 Mean 2.935 standard deviation 0.912 Sig., IBL vs. non-IBL Sig., IBL vs. non IBL within gender * Effect size for IBL intervention Average grade in subsequent required courses N 46 Mean 3.022 standard deviation 0.794 Average grade in subsequent elective courses N 17 Mean 3.052 standard deviation 1.064 Sig., men vs. women Average grade in subsequent IBL courses N 3 Mean 3.419 standard deviation 0.461 Table 2 (continued) Notes on Table 2: Average grades are reported on a 0-4 point scale. Statistically significant differences between subgroups are indicated in the marked columns: *** p<0.001; ** p <0.01; * p<0.05. Rows are included only for comparisons where at least one statistically significant difference was detected. Effect sizes are reported for the IBL intervention, only for groups or sub-groups where a significant difference was detected The IBL status of the target course is the primary independent variable by which grades and course-taking are compared, using the defined variables. We first describe the results for all IBL vs. non-IBL students, then disaggregate results by gender and by prior achievement level. Table 2 and Figure 1 show the grades and course-taking patterns of all students who had taken IBL or non-IBL sections of the target courses. In all three courses, L1, L2, and G1, IBL students grades were as good or better than their non-IBL peers (Figure 1a). In two cases, IBL students grades were statistically significantly better. When number of courses is examined (Figure 1b), little difference is seen for the more advanced students in L1 and L2. Both IBL groups took modestly (but not significantly) fewer elective courses. Students who took L1 in IBL format were more likely to opt for a second IBL course, while students in L2 had little opportunity to take additional IBL courses. Among the first-year students in course G1, however, IBL students pursued more math courses, especially IBL courses (of which three more were available). The difference in elective course count is large, though not significant due to the small sample and high variance. Because the IBL and non-IBL groups were well matched by major, the difference in elective choice is not due to differing requirements. The difference in students pursuit of IBL courses is significant. Prior results from immediate post-course surveys had flagged gender as salient in students response to IBL vs. non-IBL classes (Laursen et al., 2013b). Therefore we examined whether gender differences persisted into later courses (Figure 2, Table 2). For grades Table 3 Estimated marginal means for grades and course-taking subsequent to an IBL or non-IBL mathematics course, by prior achievement level, for one course Table 3 (continued) Sig., IBL vs. non-IBL Notes on Table 3: Average grades are reported on a 0-4 point scale. Categories for prior achievement are based on mathematics grades for courses prior to Course L1: Low = GPA 2.5, Medium = 2.5 < GPA 3.4; High = GPA > 3.4. Statistically significant differences between subgroups are indicated in the marked columns: *** p<0.001; ** p <0.01; * p<0.05. Rows are included only for comparisons where at least one significant difference was detected Fig. 1 Grades and number of subsequent courses for all students after taking IBL and non-IBL sections of courses L1, L2, and G1. (a) Average grades for next term and all subsequent required, elective and IBL courses. (b) Average number of courses taken, elective and IBL courses. Group difference is statistically significant: * at p<0.05, ** at p<0.01, *** at p<0.001 subsequent to the target course (Figure 2a), few significant differences were detected. In the upper-division courses L1 and L2, there was a minor but not significant pattern of IBL students (both men and women) out-performing non-IBL students of their own gender. For course G1, this pattern continued and is significant in one case, mens next-term grades. Comparing student grades by gender, some differences are large, at several tenths of a grade point, though not statistically significant. In the advanced course L2, women in both IBL and non-IBL groups tended to outperform their male classmates. For the mid-level course L1, there was little difference in mens and womens subsequent grades, while men in the early course G1 tended to outperform women in their own (IBL or non-IBL) sections. Fig. 2 Grades and number of subsequent courses for men and women after taking IBL and non-IBL sections of courses L1, L2, and G1. (a) Average grades for next term and all subsequent required, elective and IBL courses. (b) Average number of courses taken, elective and IBL courses. Subgroup difference is statistically significant: * at p<0.05, ** at p<0.01, *** at p<0.001 Portraying numbers of courses by gender and IBL status, Figure 2b shows that both male and female IBL students pursued further IBL courses at higher rates. Significant differences versus non-IBL students are found for both L1 (men) and G1 (men and women). IBL women from G1 persisted significantly longer, taking on average a full elective course beyond their non-IBL female peers. This pattern is not mirrored in the more advanced courses L1 and L2. By Achievement Group In interviews conducted with instructors as part of the broader study, instructors proposed that IBL would particularly benefit students with weaker academic backgrounds. The strongest students, they felt, would enjoy the challenge of IBL but would succeed in both IBL and non-IBL settings. Based on these hypotheses, we disaggregated the data for IBL and non-IBL students by prior mathematics achievement level. We present results from L1 only. L2 students took too few subsequent courses to support further division of the sample, and this analysis held no meaning for G1 where all students were high achievers. For L1, we empirically divided students into three achievement subgroups: low, GPA < 2.5; medium, GPA 2.5 to 3.4; and high, GPA > 3.4, taking care to match the underlying distributions for IBL and non-IBL samples. Table 3 and Figure 3 show the average subsequent grades of L1 students, disaggregated by IBL status and achievement level. Stairstep patterns in Figure 3a show that students prior grades tend to predict their later grades. Medium and high achievers who took L1 as an IBL course earned subsequent grades similar to those of their non-IBL peers, but low achievers from IBL sections earned consistently higher grades than their low-achieving peers in non-IBL sections. These differences were statistically significant for required and IBL courses, and large enough to hold academic meaning. For example, IBL low achievers averaged a high C+ (2.41) for subsequent required courses, while non-IBL low achievers averaged below a C (1.95). Figure 3b shows the L1 data plotted as the average difference in grades before and after the target course. Most students grades fell by several tenths of a grade point as they moved from computationally oriented, lower-division courses to more abstract, proof-based courses. Especially students with previously high grades did not maintain an A average as courses became more challenging. The contrasting pattern for IBL low achievers is thus striking: the grades of these students improved in later courses, unlike both their non-IBL peers and their higher-achieving classmates. The improvements were sizable and persistent: for instance, grades improved by about 0.3 grade points in the next term, and even more in further IBL courses. Only two of the group differences were statistically significant, but, as Table 4 shows, some of the effect sizes for the IBL intervention are substantial. Medium to large effect sizes (Cohen, 1988) are observed not only for the increase in IBL low achievers grades, but also for the lesser pre-to-post decline in grades of IBL medium- and highachievers versus non-IBL peers. We found few differences in course-taking by achievement level (Table 3). There was one statistically significant difference: high achievers who had taken an IBL section of L1 took more IBL courses than did their non-IBL peers. This finding matches instructors expectations that high achievers would find the IBL method stimulating. Overall, the effect of IBL on students subsequent grades and course-taking was modest when comparing IBL and non-IBL students in their entirety. Certainly no harm was done; Fig. 3 Grades for previously low, medium and high-achieving students after taking IBL and non-IBL sections of course L1. (a) Average grades for next term and all subsequent required, elective and IBL courses. (b) Average change in grades from all previous math courses to next term and all subsequent required, elective and IBL courses. Subgroup difference is statistically significant: * at p<0.05, ** at p<0.01, *** at p<0.001. In (b), significance of the pre-to-post-target grade change is marked on the bar; subgroup differences are marked outside the bars with brackets IBL students succeeded at least as well as their peers in later courses. This result challenges instructors common concern that material omitted to accommodate the slower pace of IBL courses may hinder student success in later courses (Yoshinobu & Jones, 2012). IBL students also tended to take additional IBL courses if available. While this study controlled for differences in prior mathematics background and achievement, other factors also affect students choice of an IBL or non-IBL section: learning beliefs, professor choice, peer influences, and even the time of day the section is offered. Our analyses may not fully Table 4 Effect size for IBL intervention on grades and course-taking subsequent to an IBL or non-IBL mathematics course Average of all subsequent math grades Average grade in the next term Average grade in subsequent required courses Average grade in subsequent elective courses Average grade in subsequent IBL courses Prior achievement group Notes on Table 4: Categories for prior achievement are based on mathematics grades for courses prior to Course L1: Low = GPA 2.5, Medium = 2.5 < GPA 3.4; High = GPA > 3.4. Effect sizes are calculated using the pretest/posttest/control design of Morris (2008) separate pre-selection from causal effects linking pursuit of further IBL courses to a good IBL experience. Positive effects of IBL on students pursuit of further mathematics courses were general to both men and women. These effects are detected among courses where student choice may be most apparent, electives and courses taught with IBL methods. The results by gender are particularly interesting given our findings from immediate post-course survey data: in non-IBL courses, women reported significantly lower learning gains than did men (Laursen et al., 2011, 2013a, b). This gender gap persisted across several types of intellectual (e.g., conceptual learning, problem-solving) and affective (confidence, interest) gains, though there were no actual differences in mens and womens grades. That is, women in nonIBL courses succeeded at similar rates to men, but reported less mastery and lower confidence at the end of the course. The present analysis shows that non-IBL women also persisted in mathematics at lower rates. In IBL courses, however, women reported similar intellectual and affective gains to men on surveys (Laursen et al., 2011), and their grades were no different. This analysis indicates that IBL women were also more likely to persist in mathematics. Enhanced persistence was apparent following G1, a course early in the curriculum, while after courses L1 and L2, such effects were less detectable as students had fewer terms left in which to adjust their major or course choices. Moreover, IBL experiences may matter more earlier in undergraduates careers (Watkins & Mazur, 2013) as also suggested by the higher gains reported by first- and second-year IBL students vs. upperclassmen (Laursen et al., 2011). Womens apparent grade improvement relative to their male peers from lower-division to advanced courses may suggest that women who persist to advanced courses are high achievers who also have high tolerance for their minority status. Disaggregated by prior achievement, differences in students grades and course-taking patterns became apparent. Taking an IBL course did not erase achievement differences among students, but did flatten them. In non-IBL courses, initial patterns of achievement difference were preserved; previously low-achieving students gained no ground. Figure 3a compares students to each other, while Figure 3b compares students to their own prior performance. Low achievers performance was boosted after taking an IBL course, relative both to their own previous performance and to non-IBL peers. Differences of 0.3-0.5 grade points are meaningful to students future academic options. The differing impact of IBL on women and low achievers shows that the intervention functions differently for these two groups. For women, the impact of IBL appears to be primarily affective; it is not permanent. IBL courses offer features that are known to be effective for women, including collaborative work (Springer, Stanne & Donovan, 1999), problem-solving, and communication (Du & Kolmos, 2009) and that may enhance womens sense of belonging to the discipline (Good, Rattan & Dweck, 2012). Public sharing and critique of student work may serve as vicarious experiences that enhance self-efficacy (van Dinther, Dochy & Segers, 2011) and link effort, rather than innate talent, to mathematical success (Good, Rattan & Dweck, 2012). For lowachieving students, however, the effect is longer-lasting. We propose that IBL experiences promote what one student called fruitful struggle, thereby strengthening transferable problem-solving strategies and study habits. For students who do not already have these skills, this is a powerful and lasting impact (Hassi & Laursen, 2013). This study also yields some methodological insight. Overall, grades and course-taking choices are blunt instruments for detecting the impact of an educational intervention. Because these outcomes and their meaning varied importantly by student sub-group, disaggregating results was essentialbut necessitated large samples. Comparing patterns in results across three courses yielded insight about the relative impact of IBL experiences on students at different academic stages. The methods used are entirely general and not specific to mathematics courses. The utility of academic records analysis was understandably sensitive to the nature and timing of the target course. Effects on subsequent grades and course-taking were most easily detected in courses earlier in the curriculum. Self- and institutional selection required the use of stringent controls that in turn required large samples. Variation by prior achievement could not be studied in the G1 course because all students were strongly prepared. Results could be rigorously compared only within a single course, not across courses or institutions. Finally, the analysis required substantial up-front work to gather institutional records, transform data, and define and compute standardized variables. In sum, academic records analysis is not a tool to be applied lightly, yet techniques like these may yield insight for studies of multi-course or multisite educational reform in cases where these design constraints can be accommodated. College instructors using student-centered methods in the classroom are often called upon to provide evidence in support of the educational benefits of their approachan irony, given that traditional lecture approaches have seldom undergone similar evidence-based scrutiny. Our study indicates that the benefits of active learning experiences may be lasting and significant for some student groups, with no harm done to others. Importantly, covering less material in inquiry-based sections had no negative effect on students later performance in the major. Evidence for increased persistence is seen among the high-achieving students whom many faculty members would most like to recruit and retain in their department. Thus these results should be useful to instructors seeking evidence to persuade colleagues and students of the value of their approach. This work raises many interesting questions for future studies. The differential benefits of inquiry learning experiences for low-achieving students highlights their potential to help overcome historical inequities for other groups, such as students of color and first-generation college studentsgroups we could not examine in this study. Comparison of longitudinal effects will be especially interesting in cases where inquiry courses are offered to first-year students and where multiple inquiry experiences are offered within a single program or institution. While the quantitative approach reported here establishes patterns of student achievement and persistence after an active-learning course, only mixed-methods approaches can both document such effects and reveal the reasons for them. Acknowledgments The Educational Advancement Foundation supported this work. We thank Clint Coburn and Tim Archie for research assistance and Andy Cameron, Michele Keeler, and Steven Velasco for access to the data. Open Access This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.


This is a preview of a remote PDF: https://link.springer.com/content/pdf/10.1007%2Fs10755-013-9269-9.pdf

Marina Kogan, Sandra L. Laursen. Assessing Long-Term Effects of Inquiry-Based Learning: A Case Study from College Mathematics, Innovative Higher Education, 2014, 183-199, DOI: 10.1007/s10755-013-9269-9