ebook img

ERIC ED592677: Exploring the Impact of Data-Driven Tutoring Methods on Students' Demonstrative Knowledge in Logic Problem Solving PDF

2016·0.34 MB·English
by  ERIC
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview ERIC ED592677: Exploring the Impact of Data-Driven Tutoring Methods on Students' Demonstrative Knowledge in Logic Problem Solving

Exploring the Impact of Data-driven Tutoring Methods on Students’ Demonstrative Knowledge in Logic Problem Solving Behrooz Mostafavi Tiffany Barnes NorthCarolinaStateUniversity NorthCarolinaStateUniversity Raleigh,NC27695 Raleigh,NC27695 [email protected] [email protected] ABSTRACT specificmethodsofproblemassignmentandcombinationof We have been incrementally adding data-driven methods hintsandworkedexamplesmayhaveimpactedstudentper- intotheDeepThoughtlogictutorforthepurposeofcreating formanceonrelatedquestionsonthecoursemidtermexam. afullydata-drivenintelligenttutoringsystem. Ourprevious research has shown that the addition of data-driven hints, In this paper we compare two classrooms of students us- workedexamples,andproblemassignmentcanimprovestu- ingdifferenttestconditionsofDeepThought,withdifferent dentperformance and retention in the tutor. In this study, combinations of problem assignment, hints and worked ex- weinvestigatehowtheadditionofthesemethodsaffectsstu- amples. Students’knowledgeoflogicwereevaluatedintwo dents’demonstrativeknowledgeoflogicproofsolvingusing problemsonamid-termexam,andthesescoreswereusedto theirpost-tutorexaminationscores. Wehaveuseddatacol- differentiatehighandlowproficiencystudentsforouranaly- lectedfromthreetestconditionswithdifferentcombinations sis. Theresultsfromouranalysisshowthathighperforming of our data-driven additions to determine which methods students benefit most from problem-solving opportunities, are most beneficial to students who demonstrate higher or whilelowperformingstudentsbenefitmostfromproblemas- lower knowledge of the subject matter. Our results show signment based on proficiency profiling, comparing current that students who are assigned problems based on profil- students to prior exemplary students with similar behav- ing proficiency compared to prior exemplary students with ior. Weconcludethattheuseofproficiencyprofilingisthe similar problem-solving behavior show higher examination mosteffectivemethodforincreasingretentionandreducing scores overall, and the use of proficiency profiling increases timespentintheDeepThoughttutor,andresultinhigher retentionandreducestheamountoftimetakenin-tutorfor overallexaminationscores. Theresultsfromthisstudyalso lower performing students in particular. The results from helpdifferentiatethebehaviorofhigherandlowerperform- thisstudyalsohelpsdifferentiatethebehaviorofhigherand ing students in tutor, allowing for quicker interventions for lowerperformingstudentsintutor,whichcanallowquicker lowerproficiencystudentswhoneedadditionalinstructional interventionsforlowerproficiencystudents. support. Keywords 2. RELATEDWORK Data-driven Methods, Proficiency Profiling, Tutoring Sys- Koedingeretal.[6]summarizedthegeneralprocessofintelli- tems genttutoringsystems: thesystemselectsanactivityforthe student, evaluates each student action, suggest a course of 1. INTRODUCTION action (either via hints, worked examples, or another form offeedback),andfinallyupdatesthesystem’sevaluationof Wehavebeenincrementallyaddingdata-drivenmethodsfor thestudent’sskills. Aneffectivetutorshouldadaptinstruc- problem assignment[9, 10], hint generation[3], and worked tion according to the student’s current knowledge level [1]. examples[11] to the Deep Thought logic tutor to create a However,inordertomakeinstructionaldecisions,mostITSs fully data-driven tutoring system. While we have observed either use fixed pedagogical policies providing little adapt- improvements in student retention and tutor scores with ability,orexpert-authoredpedagogicalrulesbasedonexist- eachoftheseadditions,wehavenotstudiedthedifferencein ing instructional practices [1, 14]. Intelligent tutoring sys- post-tutor examinations when these methods are combined tems with data-driven methods can be more adaptive by indifferenttestconditions. Weseektounderstandhowthe leveraging previous student data in order to complete one or more of these steps. Data-driven approaches to mak- ing effective pedagogical decisions – in particular selecting problems,whentoapplyworkedexamples,andthetypeof hintorfeedbacktoprovide–wouldmostlybypasstheneed for expert involvement in creating and improving the effec- tiveness of ITSs. In practice, incorporating student data has been shown to increase learning efficiency and predict student behavior. This, in particular is why we use data- drivenknowledgetracing(DKT)ofruleapplicationswithin Proceedings of the 9th International Conference on Educational Data Mining 460 the Deep Thought logic tutor to facilitate profiling of stu- hierarchical clustering. Expert weighting was replaced by dents’proficiency. PCA of the frequency of the rules used for each exemplar for each level, accounting for 95% variance of the results. Intheremainderofthissection,wedescribetheDeepThought Foreachrule,itsPCAcoefficientisthenewweightforthat logictutorandthedata-drivenadditionsimplemented. We rulescore. Whenanewstudentusesthetutor,thestudent’s then describe the system and data used to evaluate the ef- rule scores are calculated throughout the level. At the end fectiveness of these data-driven methods in Deep Thought. ofeachlevel,theDDPPexamineseachstudent’sindividual Afterreportingtheresultsofthisevaluation,wediscussthe rulescoreandassignsittoaclusterforthatrule. TheDDPP implications for future design decisions in the tutor, and thenfindswhichclustersthescoresforthemostimportant presentourconclusions. rules fall into for that level (based onthe same PCA based weighting),andthenclassifiesthatstudentintoatypebased on the set of clusters the student matches. Finally the sys- 2.1 TheDeepThoughtTutor temassignsthestudenttoaproficiencytrackbasedondata Wehavebeenexaminingthepotentialfordata-drivenmeth- fromthematchingtypeofexemplars,andhowthoseexem- ods to improve learning gains in a complex problem solv- plars were placed in the next level. The more exemplars ingdomainbyincrementallyaugmentingtheDeepThought wehaveofagiventype,thestrongerthepredictionwecan logic tutor. Deep Thought is a tutor for graphically con- make for a new student. In the event that a new student structingpropositionallogicproofs. DeepThoughtpresents doesn’t match an existing type in the exemplar data, the proof problems consisting of logical premises and a conclu- student’s proficiency is calculated using the average scores, sion to be derived using logical axioms. Deep Thought is asintheoriginalDDMLsystem. divided into 6 levels of logic proof problems. In previous work with the Deep Thought logic tutor, we have been im- Providing hints to students in the course of an intelligent plementingdata-drivenmethodsforseveraloftheintelligent tutor as a possible form of step-based feedback has the po- tutorsteps. Weimplementedadata-drivenmasterylearning tential to increase learning gains. Razzaq, Leena, and Hef- system(DDML)totrackstudentactionsandassignappro- fernan [12] found that learning gains increased for students priateproblemsbasedonthestudent’scurrentlevelofprofi- givenon-demandhintsincomparisontostudentswhowere ciency[9]. Theproblemsetwassplitintotwotracks: ahigh provided hints proactively. In Deep Thought, the hint sys- proficiencytrackandalowproficiencytrackforLevels2–6, tem used is called Hint Factory. Hint Factory is an auto- withLevel1containingacommonsetofproblemsforinitial matic data-driven hint generator that converts an interac- track assignment. We tracked student actions throughout tionnetworkgraphofstudenttracebehaviorintoaMarkov theirtimeinthetutor,andinparticulartheirapplicationof decision process (MDP) to automatically select on-demand logicalrulestoconstructlogicproofs. Basedontheircorrect hints for students upon request, based on their individual orincorrectapplicationoflogicalrules,theDDMLupdated performanceonspecificproblems. TheMDPisdata-driven, asetofrulescores,onescoreforeachlogicalrule. Attheend using actions logs from previous Deep Thought use in the of each level, the students’ rule scores were weighted based classroom to assign weight to proof-state actions based on on expert-determined priorities; rules deemed by experts whetherornotthatactionultimatelyledtosuccessfulcom- to be of high importance to solving the problems in that pletion of the proof. These hints help students solve prob- level were weighted higher than rules that were not. These lems by suggesting what step should be taken next on a weightedscoresweresummedtogether,andcomparedtothe multi-step problem. Hint Factory has been implemented average rule scores in the previous semester’s data; based in the Deep Thought logic tutor to automatically deliver onthiscomparison,studentswereassignedtothehigheror context-specifichintstostudentsduringproblem-solving[4]. lower proficiency path. We tested Deep Thought with the In a previous study Hint Factory was shown to provide DDML incorporated and found students completed, on av- context-specific hints over 80% of the time [3]. In a pilot erage,79%ofallsixlevelsinthetutorassignment. Student study,Barnes&StamperfoundthatHintFactorycanpro- retentionratewas55%. Thiswasanimprovementoverthe vide sufficient, correct, and appropriate hints for the Deep non-DDMLversionofDeepThought(61%tutorcompletion Thought Logic tutor and help students to solve more logic onaverage,and31%retentionrate). proof problems in the same span of time [4]. However, we currently cannot determine the effect hints would have in Welaterincorporatedadata-drivenproficiencyprofiler(the addition to the DDML or DDPP; so far, students using ei- DDPP) to replace the expert-determined priorities[10][8]. ther of those versions of Deep Thought did not use hints The DDPP is a system that calculates student proficiency oftenenoughforanymeaningfulanalysis. at the end of each level in Deep Thought based on how a given student performs in comparison to exemplars who Addingworkedexamplesasasupplementtotraditionalprob- employedsimilarproblemsolvingstrategies,withrulescores lemsolvingcanalsobebeneficial[2,13]. HilbertandRenkl weighted as determined through principal component anal- [5] found that improved learning outcomes occurred when ysis(PCA).Basedonhowsimilarexemplarystudentswere providing worked examples with a prompt, and proposed assignedinsubsequentlevels,theDDPPcandeterminethe thatthiswasduetoallowingthestudentstohaveagreater best proficiency level for a new student. In contrast to the cognitive load at once. McLaren and Isotani [7] compared DDML system previously employed, this proficiency calcu- three tutors using all worked examples, all traditional un- lation and rule weighting is entirely data-driven, with no guidedproblemsolving,andamixofworkedexamplesand expertinvolvement. problemsolving. Eachgroupachievedsimilarlearninggains, but the students who were given all worked examples re- Wedeterminedsimilarproblemsolvingstrategiesamongthe quired less time to achieve those gains. We added worked exemplarsbyclusteringtheexemplars’rulescoresbasedon Proceedings of the 9th International Conference on Educational Data Mining 461 examples to the version of Deep Thought with the DDML by performance on the post-test and by the predominant incorporated[11]. Workedexamplesweregeneratedbasedon track level in Deep Thought. The post-test was a set of previousbeststudentsolutions,andprocedurallyannotated. two proofs students had to solve on paper for a midterm Theywerepresentedtostudentsrandomlyonaper-problem exam. Thesequestionswerehand-gradedwithpartialcredit basis, based on the number of problems they had solved in given based on the percentage of the proofs completed and that level already. We found that student retention overall pointstakenoffformisapplicationofrulesandskippingnon- was 90%, and students completed 94% of the tutor on av- trivial rules. We considered two performance levels: post- erage. Thispercentagewassignificantlyhigherthanthatof test scores greater than or equal to 80% (AB), or less than theDDMLalone. 80% (CDF). The post-test scores mark the final evaluation of students’ ability to solve proof problems, and occurs im- mediately following the Deep Thought tutor homework as- 3. METHODS signment. Deep Thought was used as a mandatory homework assign- mentbystudentsinanundergraduate“discretemathemat- Theseconddimensionwestudiedwastheproportionofhigh ics for computer scientists”course in Fall 2015 and Spring tolowproficiencytracklevelsthestudentscompleted. Stu- 2016. Students in the two semesters were taught by differ- dents who were assigned to the high proficiency track in a ent instructors. Students were assigned Levels 1–6 of Deep levelhadtheabilitytofinishoneitherthehighorlowprofi- Thoughtforfullcredit,withpartialcreditawardedpropor- ciencytrackdependingonthenumberofproblemsskipped tionaltothenumberoflevelscompleted. Forthisstudy,we within that level. Students who completed more levels on comparethedatafromthreeDeepThoughttestconditions thehightrackthanthelowtrackweremarkedashightrack used across the two semesters to differentiate the effect of students,andstudentswhocompletedmorelevelsonthelow ourdata-drivenmethodsonstudentperformance. trackthanthehightrackweremarkedaslowtrackstudents. Thetrackassignmentsindicatethenumberandcomplexity The first group evaluated for this study were assigned only of problems students received, with the low track having problem-solving opportunities (PS group, n = 26). The moreproblemsoflowercomplexity,andthehightrackhav- problemassignmentsystemusedwastheDDMLsystemde- ing fewer problems of higher complexity. The tracks were scribedintheprevioussection,wherestudentswereassessed designed so that students would have a similar number of betweenlevelsandplacedoneitherahighorlowproficiency ruleapplicationsacrossthetracks,eventhoughthenumber track in the next level. This group of students were taken ofproblemsdiffers. Typically,thelowtrackhasthreeprob- onlyfromtheFall2015semester,asthereexistednoequiv- lemswithexpertsolutionsusing5ruleapplications,andthe alenttestconditioninSpring2016. high track has 2 problems with expert solutions using 7 – 8 rule applications - meaning that both tracks minimally Thesecondgroupofstudentswererandomlyassignedeither required about 15 total rule applications (though students problem-solving opportunities or worked examples of the typicallyusedmore). same problems within each level (PS/WE group, n=179), withthenumberofproblem-solvingopportunitiescontrolled Inadditiontopost-testandpredominanttracklevel,weex- to match the number of problems solved by the PS group. aminedtotaltimeintutor,averagetimespentperproblem, Like the PS group, the PS/WE group were assigned profi- percentage of correct rule applications out of all rule appli- ciencytracksusingtheDDML.However,becauseindividual cations, andthetotalnumberofruleapplications. Wealso ruleapplicationscoreswereupdatedateachstepinworked lookedatancillarybehaviors(hintusage,skippedproblems, examplesasifastudenthadappliedthatruleinwhileprob- andreferencerequests)thatcoulddifferentiatehighandlow lemsolving,moststudentswereconsistentlyassignedtothe performing students. We compared these metrics to better high track in most levels, and were only assigned the low understandtheimpactofworkedexamples,hints,anddata- trackwhentheirindividualperformancewasbelowsatisfac- driven track selection on student performance. The results tory. This group of students were taken from both the Fall ofthisdescriptiveanalysisarepresentedinthenextsection. 2015andSpring2016semesters. Thethirdgroupofstudentswererandomlyassignedproblem- 4. RESULTS solvingopportunitiesorworkedexamplesinthesameman- Table1displaysthepercentageofABstudentsineachofthe ner as the PS/WE group, but with the DDPP method as- PS, PS/WE, and DDPP groups for all students, as well as signingproficiencytracksinsteadoftheDDML,wherestu- students who completed the majority of the tutor in either dentswereassignedthesameproficiencytrackaspriorstu- thehighorlowtracks. Table1alsodisplaysthepercentage dents who most closely matched their rule application be- ofstudentsineachgroupandeachtrackwhodroppedoutof havior (DDPP group, n = 61). This group of students thetutorbeforefullcompletion,asthisisoneofthemetrics were also taken from both the Fall 2015 and Spring 2016 we have used to judge the effectiveness of our data-driven semesters. Students in all three groups had access to on- methods. In our previous work using the same version of demandhints. Deep Thought, we found that students completed 94% of the tutor on average, with a retention rate of 90%. The All students were evaluated using two proof problem ques- averagepercenttutorcompletionforthegroupsinthisstudy tions as part of a mid-term examination, which was used wereconsistentwiththesenumbers(PS:95%,PS/WE:93%, as a post-test for this study. Students performance in the DDPP:94%). post-testforbothFall2015andSpring2016weregradedby the same teaching assistant, ensuring consistent evaluation Thefirstinterestingresultofnoteisthatthepercentageof across all results. Students were separated for evaluation students who performed better on the post-test was higher Proceedings of the 9th International Conference on Educational Data Mining 462 AsshowninTable2,amongABstudentsinallthreegroups, Table1: PercentageofABStudentsandPercentage the total time spent in tutor appears similar, although the of dropped students in the PS, PS/WE, and DDPP mean time for high-track students was lower for DDPP (M groups. =3.95hr,SD=6.21hr)comparedtoPS/WE(M =4.46hr, Condition ALL High Track Low Track SD = 9.13hr) and PS (M = 6.66hr, SD = 9.91hr). The n % AB Students mean time for low-track students was lower for PS (M = PS 26 65.38 71.43 63.16 4.63hr, SD = 9.55hr) and DDPP (M = 5.48hr, SD = PS/WE 179 49.72 52.35 36.67 5.42hr) than the PS/WE (M = 7.74, SD = 9.76). The DDPP 61 63.93 66.67 61.76 means of average problem time, percentage of correct rule n % Dropped Students applications, and number of rule applications were consis- PS 26 3.85 0.00 5.26 tent with the median values presented in Table 2 across all PS/WE 179 11.73 6.71 36.67 three groups. Note that low-track students in the PS/WE DDPP 61 9.84 11.11 8.82 groups had the lowest percentage of correct rule applica- tions,andthehighestnumberoftotalruleapplicationsamong allthegroups. Thismeanstheyaredoingmorework,buta lowerpercentageofitiscorrect. forforthePS(65%)andDDPP(64%)groupsthanforthe PS/WE group (50%), across all the students, as well as As shown in Table 2, among CDF students in all three within the high and low track groups. In the PS group, groups, the total time spent in tutor is dramatically dif- students who completed more levels on the high track dis- ferent, with PS spending 3 to 4 times as long in the tutor playedahigheroverallproficiencyofthesubjectmatterthan than PS/WE and DDPP groups. This ratio is also similar thosewhofinishedmoreoftenonthelowtrack(71%vs63%, in the average problem time for high and low track stu- respectively),asdidstudentsinthePS/WEgroup(52%vs dents, and the number of total rule applications for high 37%). track students. Therefore, while problem-solving only (PS) may have a slightly higher overall success rate in helping However, students in the DDPP group showed a consis- students learn proof problem solving and remain in the tu- tent level of proficiency regardless of the tracks completed torthantheDDPPstudents,forstudentswhoarelesspre- (66% vs 61%), which makes sense considering that these pared, PS results in a much higher time spent in the tu- students were matched to previous successful students who tor, with little return on the time investment. Therefore, displayedsimilarrule-applicationbehavior,andhadamore forstudentswhohaveabettergraspofthesubjectmatter, even placement within the high and low tracks compared pure problem-solving may offer a slightly better option for to the PS group, who had even placement among tracks, gettingthroughtheassignedtutor,althoughthedifferences but within the context of their own performance compared betweenproblemsolving,problemsolvingandworkedexam- to expert-decided thresholds. The DDPP group also had ples, and proficiency profiled assigned problem solving and higherplacementcomparedtothePS/WEgroup,whowere worked examples are minimal. However, for less prepared placed on the high track much more often than not due to students, pure problem-solving opportunities offer little to theinclusionofworkedexamples. AKruskal-Wallistestfor guidestudentstohigherunderstandingofthematerial,and one-wayanalysisofvarianceshowednosignificantdifference ingeneral,theDDPPoffersamuchbetterpathtocomplet- betweengroups(p=0.22). ingthetutorinfarlesstimeforbothABandCDFstudents, givingstudentstheopportunitytoencounterallthesubject Students also had a higher retention rate in both the PS matter and have a greater chance of learning the material, (4%) and DDPP (10%) groups compared to the PS/WE resultinginhigheroverallpost-testscores. group(12%). Itisespeciallyinterestingtoseethedroprate among low track students in the PS/WE group, who had Completingthetutorassignmentisimportantforstudents; a much lower retention rate among all the students in the however,sincewewanttomakesurethatstudentsarelearn- study. Because students in the PS/WE group were more ingthematerialwell,mid-termexaminationscoresareulti- often that not placed in the high track in each level, for mately a higher gauge for learning success. Among all the studentstoenduponthelowtrackindicatesahighlevelof experimentalgroupsinthisstudy,atmost65%ofstudents problem-skipping among these students. We can conclude were performing at A or B grade level on the mid-term ex- that low performing students who are not intelligently as- amination. WewouldliketoincreasethispercentageofAB signedproblemsbasedontheirproblem-solvingperformance students,sothequestionatthispointis: Isitpossibleforus appeartogainlittlefromworkedexamples. to predict low exam scores based on in-tutor data for early intervention? While it may be tempting to declare problem-solving op- portunitieswithnoworkedexamplesasthebestperforming We first look at the differences between AB and CDF stu- pedagogical choice among the three groups based on these dentsinTable2,withtheassumptionthattheDDPPmethod numbers alone, a look into additional performance metrics offers the best overall chance of success for students. For gives some more insight. Table 2 presents the amount of hightrackstudents,totaltutortime,averageproblemtime, timespentintutorandoneachproblem,aswellastheper- percentage of correct rule applications, and total rule ap- centageofcorrectandtotalruleapplicationsforeachgroup, plications are consistent between AB and CDF students. separatedbytrack. Thenumberspresentedarethemedian However,forlowtrackstudents,averageproblemtime,per- valuesforeachmetric,sincethedistributionsofscoreswere centage of correct rule applications, and total rule applica- highly skewed and non-normal, and none of the differences tionsshowahigherdifference. CDFstudentsspenttwiceas weresignificantduetolowsamplesizewithineachsubgroup. Proceedings of the 9th International Conference on Educational Data Mining 463 Table 2: Total Time, Average Problem Time, Percentage of Correct Rule Applications, and Total Rule Applications for AB and CDF students in the PS, PS/WE, and DDPP groups, separated by High and Low Track. The numbers listed are all median values. AB STUDENTS CDF STUDENTS PS PS/WE DDPP PS PS/WE DDPP HIGH TRACK n 5 78 18 n 2 71 9 Total Tutor Time (hr) 2.47 2.37 2.80 12.8 3.75 3.17 Average Problem Time (min) 9.89 11.1 12.1 52.3 18.4 16.0 % Correct Rule Applications 60.8 63.5 58.5 64.1 56.9 62.3 Total Rule Applications 258 214 203 471 255 204 LOW TRACK n 12 11 21 n 7 19 13 Total Tutor Time (hr) 1.80 3.33 3.67 17.2 5.96 4.98 Average Problem Time (min) 6.76 15.2 15.0 60.1 25.0 30.4 % Correct Rule Applications 68.8 45.5 57.0 48.7 45.7 47.0 Total Rule Applications 201 404 291 382 394 389 longonaverageperproblemthanABstudents,andapplied rulescorrectlylessthanhalfofthetime,whileABstudents applied rules more than half of the time. CDF students also attempted applying rules 25% more overall than AB students. 5. CONCLUSION In this paper we compared two classrooms of students us- SincetheperformancedifferencesbetweenABandCDFstu- ingdifferenttestconditionsofDeepThought,withdifferent dents are not as apparent for high track students, we look combinationsofproblemassignment(DDMLorDDPP)and atancillarytutorbehaviortomakeabetterdistinction. Ta- the addition of worked examples, for the purpose of under- ble 3 shows the number of requested hints, the number of standing how the specific methods of problem assignment skippedproblems,andthenumberofrulereferencerequests and combination of hints and worked examples affect high (descriptions of logic rule operations) made by students in and low performing students, as evaluated using mid-term all groups. For the DDPP group, the most apparent dif- examination scores. We found that for higher proficiency ference among AB and CDF students are the number of students who have a firmer grasp of the subject matter, hints requested, with the CDF group requesting 32 hints problem-solvingopportunitiesofferthebestchanceofcom- (M =50,SD=57)comparedto17(M =32,SD=42)for pletingthetutorinatimelymanner;however,theaddition the AB group. This difference in hints requested between ofworkedexamplesdoesnotsignificantlydetractfromthese AB and CDF students is also consistent across all groups students’ learning experience. The method of problem as- and both high and low track students. We conclude that signment (DDML or DDPP) does not have a noteworthy forhightrackstudents,wecandifferentiatebetweenhigher effectonhighstudentperformance. and lower proficiency students using hint request behavior, and for low track students, we can differentiate higher and Forlowerproficiencystudents,wefoundthatproblem-solving lower proficiency students using the amount of time spent opportunitiesalonewithDDMLproblemassignmentoffered on average per problem and the percentage of correct rule little to guide students to higher understanding of the ma- applications. Thisallowsthepossibilityofmakinganinter- terial, and greatly extended the amount of time students vention during a student’s progress through Deep Thought spent in the tutor with little learning benefit. The addi- in the case that a student requires additional feedback or tion of worked examples helped these students get through aid from an instructor due to a lesser understanding of the the tutor faster, however these students had a lower reten- subjectmatter. tion rate than any other students and lower examination scores. We conclude from these results that updating our data-driven skill estimates equally for viewing or applying rules resulted in students being assigned to the high-track Table 3: Number of Hints, number of Skips, and whentheywerenotpreparedtosolveharderproblems. With numberofRuleReferencerequestsforABandCDF proficiencyprofiling–matchingstudentstopreviouslysuc- studentsinthePS,PS/WE,andDDPPgroups,sep- cessfulstudentsandthepathstheytakethroughthetutor– arated by High and Low Track. The numbers listed wecanreducetheamountoftimespentintutor,increasere- are all median values. tention, andmakebetteruse of workedexamplesby giving PS PS/WE DDPP them alongside problems that better match an individual HIGH AB CDF AB CDF AB CDF student’s proficiency level. This results in similar perfor- # Hint 95 166 12 26 17 32 mance to problem solving alone in terms of retention and # Skip 5 16 1 1 0 2 knowledge gained, but with a lot less time spent in the tu- # Ref 151 168 76 145 111 92 tor for lower-proficiency students. We conclude that our LOW AB CDF AB CDF AB CDF DDPP method offers the best overall possibility of success # Hint 30 104 31 44 19 26 forstudentscompletingtheDeepThoughttutorinatimely #Skip 1 0 30 24 3 15 manner, learning the subject matter, and performing well # Ref 77 224 60 271 55 109 onpost-tutorexaminations. Proceedings of the 9th International Conference on Educational Data Mining 464 6. ACKNOWLEDGEMENTS This material is based on work supported by the National ScienceFoundationunderGrants1432156and0845997. 7. REFERENCES [1] J.R.Anderson,A.T.Corbett,K.R.Koedinger,and R.Pelletier.CognitiveTutors: LessonsLearned.The journal of the learning sciences,4(2):167–207,1995. [2] R.K.Atkinson,S.J.Derry,A.Renkl,and D.Wortham.Learningfromexamples: Instructional principlesfromtheworkedexamplesresearch.Review of educational research,70(2):181–214,2000. [3] T.Barnes,J.Stamper,L.Lehmann,andM.J.Croy. Apilotstudyonlogicprooftutoringusinghints generatedfromhistoricalstudentdata.InProceedings of the 1st International Conference on Educational Data Mining (EDM 2008),pages197–201,2008. [4] T.Barnes,J.Stamper,L.Lehmann,andM.J.Croy. Apilotstudyonlogicprooftutoringusinghints generatedfromhistoricalstudentdata.InProceedings of the 1st International Conference on Educational Data Mining (EDM 2008),pages197–201,2008. [5] T.S.HilbertandA.Renkl.Learninghowtousea computer-basedconcept-mappingtool: Self-explaining exampleshelps.Computers in Human Behavior, 25(2):267–274,2009. [6] K.R.Koedinger.NewPotentialsforData-Driven IntelligentTutoringSystemDevelopmentand Optimization.AI Magazine,34(3):27–41,2013. [7] B.M.McLarenandS.Isotani.Whenisitbestto learnwithallworkedexamples? InArtificial Intelligence in Education,pages222–229,2011. [8] B.MostafaviandT.Barnes.Data-drivenProficiency Profiling-ProofofConcept.InProceedings of the 6th International Conference on Learning Analytics and Knowledge (LAK 2016). In Press.,2016. [9] B.Mostafavi,M.Eagle,andT.Barnes.Towards data-drivenmasterylearning.InProceedingsofthe5th International Conference on Learning Analytics and Knowledge (LAK 2015),pages270–274,2015. [10] B.Mostafavi,Z.Liu,andT.Barnes.Data-driven ProficiencyProfiling.InProceedings of the 8th International Conference on Educational Data Mining (EDM 2015),pages335–341,2015. [11] B.Mostafavi,G.Zhou,C.Lynch,M.Chi,and T.Barnes.Data-drivenWorkedExamplesImprove RetentionandCompletioninaLogicTutor.In Proceedings of the 17th International Conference on Artificial Intelligence in Education (AIED 2015), pages726–729,2015. [12] L.RazzaqandN.T.Heffernan.Hints: isitbetterto giveorwaittobeasked? InProceedings of the 10th International Conference on Intelligent Tutoring Systems (ITS2010) Part 1,pages349–358.Springer, 2010. [13] J.SwellerandG.A.Cooper.Theuseofworked examplesasasubstituteforproblemsolvingin learningalgebra.Cognition and Instruction, 2(1):59–89,1985. [14] K.VanLehn.TheBehaviorofTutoringSystems. International Journal of Artificial Intelligence in Education,16(2):227–265,2006. Proceedings of the 9th International Conference on Educational Data Mining 465

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.