ebook img

ERIC ED401812: The Effects on Learning, Course Evaluation, and Team Evaluation of Changing STAD Teams at Midterm. PDF

14 Pages·1996·0.24 MB·English
by  ERIC
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview ERIC ED401812: The Effects on Learning, Course Evaluation, and Team Evaluation of Changing STAD Teams at Midterm.

DOCUMENT RESUME HE 029 660 ED 401 812 Gnagey, William J.; Potter, Kirsten I. AUTHOR The Effects on Learning, Course Evaluation, and Team TITLE Evaluation of Changing STAD Teams at Midterm. PUB DATE [96] NOTE 13p. Research/Technical (143) Reports PUB TYPE MFO1 /PCO1 Plus Postage. EDRS PRICE *Academic Achievement; *College Students; DESCRIPTORS *Cooperative Learning; Course Content; *Course Evaluation; Educational Psychology; *Grouping (Instructional Purposes); Heterogeneous Grouping; Higher Education; *Peer Evaluation; Peer Teaching; Teamwork *Student Teams Achievement Divisions IDENTIFIERS ABSTRACT This study is the seventh in a series of investigations on the effectiveness of the Student Teams Achievement Division (STAD) approach to cooperative learning among college students. The STAD approach features heterogeneous teams of four to six students who tutor each other on course materials and prepare each other for weekly quizzes. Students take the quizzes individually, but are awarded bonus points on the basis of their team's mean quiz score. A total of 196 undergraduates in four sections of an educational psychology course were divided into STAD teams of five or six students, with half of the students being placed in new teams halfway through the semester. T-tests were performed on both groups in regard to quizzes, examinations, course ratings, and team member performance appraisals. The study found that being assigned to a new team at midterm had little or no effect on learning the course subject matter, evaluating the course, or evaluating teammates. Copies of course and team member evaluation forms are included. (Contains 11 references.) (MDM) *********************************************************************** * * Reproductions supplied by EDRS are the best that can be made * * from the original document. *********************************************************************** TEAM THE EFFECTS ON LEARNING, COURSE EVALUATION, AND EVALUATION OF CHANGING STAD TEAMS AT MIDTERM William J. Gnagey & Kirsten I. Potter Illinois State University 00 BEST COPY AVAILABLE "PERMISSION TO REPRODUCE THIS U.B. DEPARTMENT OF EDUCATION MATERIAL HAS BEEN GRANTED BY Office of Educational Research and Improvement EDUCATIONAL RESOURCES INFORMATION William J. Gnagey CENTER (ERIC) is document has been reproduced as received from the person or organization originating it 0 Minor changes have been made to improve reproduction Quality. TO THE EDUCATIONAL RESOURCES Points of view or opinions stated in this docu- INFORMATION CENTER (ERIC)." ment do not necessarily represent official OERI position or policy. EVALUATION, AND TEAM THE EFFECTS ON LEARNING, COURSE EVALUATION OF CHANGING STAD TEAMS AT MIDTERM William J. Gnagey & Kirsten I. Potter Illinois State University Perspective for of investigations of the effectiveness This experiment is the seventh in a series of Achievement Divisions (STAD), one system university students of Student Teams been based on the research of Robert E. cooperative learning. These studies have public schools. primarily on STAD's use in Slavin (1983, 1989, 1990, 1991) who focused each other of 4-6 members who tutor This system features small heterogeneous teams weekly quizzes that measure and prepare each other for on the material in the course awarded bonus quizzes individually, but are chapter objectives. Students take the quiz score. points on the basis of the team's mean sophomore educational 1988), four sections of In our first experiment (Gnagey, and control divided into equivalent experimental psychology students (N=145) were taught in the STAD three week baseline period, both groups were groups. After a awarded students in the control group were format for the rest of the semester. While students in the improvement of their teams, bonus points on the basis of the mean improvement. for their own individual experimental group received bonus points the two groups, found in the achievement of Although no significant differences were classes. effective than did the control the experimental classes rated the course as more hoped that the social (Gnagey & Ostrowski, 1991), it was In our second experiment their individual be prevented by publicizing loafing of some team members could students (N=75) of educational psychology contributions to their teams. Two classes conventional experimental sections were taught in a randomly assigned to control and experimental and baseline period. Both the lecture-discussion format for a four week remainder of the the STAD approach for the control sections were taught using concerning their team's received anonymous feedback semester. The control class additional feedback experimental class received quiz performance, whereas the No significant other members of their own teams. making them privy to the scores of for the quiz, midterm, and final exam scores differences were found between the mean two sections. tried to determine the (Gnagey & Osfrowski, 1992), we In our third experiment the STAD format. awarding bonus points within differential effects of two ways of forty-student educational development classes and two Two forty-student adolescent During the first half of in parallel experiments. psychology classes served as subjects in which bonus points in the same STAD format the semester, all classes were taught quiz with the highest mean chapter members of the three teams were awarded to all chosen students in one randomly During the last half of the semester, scores. psychology class randomly chosen educational adolescent development class and one points. their team's mean improvement bonus points on the basis of were awarded For each experiment, other two classes remained the same. The award structure of the 1 3 the mean scores of the experimental and control groups were compared on seven quizzes, a final examination, and the University Course Rating Scale. In each experiment, the "mean improvement points" class outscored the "mean raw score" class on only one chapter quiz. It was concluded that the independent variable was not robust enough to materially influence either students' achievement or their evaluation of the course. In our fourth experiment (Gnagey & Navarro, 1993), we compared the effects of two methods of constructing learning activities for classes being taught in the STAD format. One hundred fifty-four students in four sections of educational psychology served as subjects. During the first half of the semester, all sections were taught in the same STAD format during which the instructor authored and assigned activities designed to assist team members in preparing each other for the weekly chapter quizzes. During the second half of the semester, two randomly chosen experimental classes prepared and carried out their own activities, while the two control sections continued to use those prepared by the instructor. Comparisons of the experimental and control group means on five chapter quizzes, the final examination, and the University Course Rating Scale revealed no significant differences on any of these measures of achievement or the course evaluation. It was concluded that after one-half a semester's experience with instructor-constructed learning activities, student teams were able to plan and carry out some that were at least as effective as those constructed by their teacher. In our fifth experiment (Gnagey & Navarro, 1994), we wanted to see if awarding bonus points for both mean quiz scores and mean improvement points would produce higher class achievement and course evaluation than either method by itself. One hundred fifty-four educational psychology students in four sections of educational psychology served as subjects in the experiment. At the beginning of the semester, all sections were taught in the usual STAD format with bonus points being awarded for correctly completing learning activities on Wednesdays, and for being in one of the four teams scoring highest on the weekly chapter quizzes. After Quiz 4, one section chosen at random began getting bonus awards for improvement scores while the other three sections continued being awarded points for high mean team quiz scores. After the midterm examination, two of these three teams were randomly assigned to other treatments. In one section, the two teams with the highest mean quiz scores and the two with the highest mean improvement scores were awarded bonus points. In the second, the four teams with the highest mean improvement points were rewarded. At the end of the semester, a final examination, and the University Course Rating Scale were administered to all sections. No significant differences were found on any of these comparisons save for two chapter quizzes. It was concluded that none of the experimental variations had a systematic effect on either learning or course evaluation. In our sixth experiment (Gnagey and Denoyer, 1995) we compared process and product methods of awarding bonus points to STAD teams. One hundred sixty-five students in four sections of educational psychology served as subjects in the 2 experiment. They were assigned by the usual registration procedures employed by the University and were mostly sophomores taking their first course in a program designed to produce high school teachers. the For the first three weeks of the semester, all four classes were taught according to usual STAD format. A General Linear Models Analysis indicated that there were no significant differences among the four sections on any of the first three quizzes. At this time, two experimental and two control groups were selected using a random number table. While the Wednesday teamwork and Friday quizzes remained the sections was graded as a same for all sections, the teamwork for the experimental project instead of being the basis for bonus points. Bonus points were awarded instead for mean quiz scores. In essence, the experimental subjects received bonus points for a learning product (quiz scores) while the control sections continued to be rewarded for a learning process (learning activities). T-tests were performed between the combined experimental and combined control group data for the remaining nine chapter quizzes, the midterm and final examinations, and the first and second administrations of the University Course Rating Scale (a course evaluation device) and the Team Member Performance Appraisal (a combined rating of the effectiveness of one's team members). Since no significant differences appeared in any of these analyses, it was concluded that teachers may award extrinsic reinforcers for either learning activities (the learning process) or quiz scores (the learning product) with the equal results in the acquisition of subject matter, the evaluation of the course, and attitudes developed among teammates. Objectives The objectives of the present experiment concern the differential effects of a one-team and a two-team approach to STAD. The following three hypotheses were tested: Students who stay with the same STAD team for the entire 1. who semester, will learn more of the subject matter than students change teams at midterm. Students who stay with the same STAD team for the entire 2. students who semester will rate the course as more effective than change teams at midterm. 3. Students who stay with the same STAD team for the entire effective than semester will rate their teammates as more students who change teams at midterm. Methods and Techniques educational psychology served as One hundred sixty-nine students in four sections of usual registration procedures subjects in the experiment. They were assigned by the taking their first course in a employed by the University and were mostly sophomores school teachers. program designed to produce high 3 During the second day of class, all students were assigned to one of eight five- or six- member teams by asking first the women and then the men to count off by eight's. Students with the same number became a team which met at a place designated by a diagram on the board. The first teamwork assignment was to learn each other's names and trade addresses and phone numbers. For the first eight weeks of the semester, all four classes were taught according to the Each Monday, the assigned chapter was introduced and usual STAD format. appropriate material was presented by way of lectures, films or videos. Each Wednesday, heterogeneous teams of 4-6 members convened to work on projects which involved the application of the text material to practical situations. Bonus points were awarded for accurate work. into Each Friday after the corrected teamwork was returned, the class was divided held a question and answer two activities. At the beginning of the hour, the instructor in the session in which students could ask for clarification of difficult concepts chapter quiz chapter. When there were no more questions, a 20-item multiple choice individually. At the end of the Friday class, the was administered to all students the appropriate chapter assignment for the following week was written on chalkboard. introduced, all During class on the following Monday before the new chapter was by student social quizzes were returned, and a list of "grades so far" was posted Wednesday, to any security numbers. Students were encouraged to write rebuttals, by STAD cycle then of the items they missed but felt they should have credit for. The began for the new chapter. administered over the first six At midterm, a 60-item multiple choice examination was administered at this chapters. The University Course Rating Scale (UCRS) was also (See Appendix 1). time as was the Team Member Performance Appraisal instrument. asked to evaluate all The UCRS is composed of twelve items on which students are and papers, textbook and other aspects of the course: clarity of objectives, projects feedback, interestingness, assigned readings, in-class activities, quizzes and exams, learned, expected level of instruction, grading procedures, grading fairness, amount performance (standards), and group activities. Ostrowski, 1992), factor analysis of the In a previous study of 158 students, (Gnagey & three principal factors (See UCRS (see appendix) using varimax rotation revealed loaded heavily on items D Table 1.). The first accounted for 31% of the variance and (amount learned), K ( (in-class activities), G (interestingness), H (instruction); J second principal factor expected level of performance), and (group work). The of the variance and loaded heavily on items A (clarity accounted for 29% of and exams), and I objectives), B (out of class papers and projects), C (text), E (quizzes 4 6 and (fairness of grades). The third factor accounted for only 11% of the variance loaded on item F (feedback). .81 using 48 The test-retest coefficient of reliability for the UCRS total score was other eight students in similar classes with one administration at midterm and the weeks later during the finals. scale (see Appendix 2) was A Team Member Performance Appraisal (TMPA) rating members of administered requiring all students to rate themselves and the other team attendance, b) contribution of their teams on the following criteria: a) team meeting feelings among team ideas, c) completion of assignments, d) promoting positive members. The ratings were members, e) encouraging the expression of other team indicating "almost never" and five done using a 5-point scale for each criterion: one indicating "almost always". eighth week, a random number table After the midterm examination at the end of the the other two sections as experimental groups while was used to designate two in the same manner as before. T- became control groups which continued to be taught experimental and combined control group tests were used to compare the combined began (See Table variables before the experimental treatment scores on all dependent the of the nine comparisons. Students in 2.) No significant differences appeared in any using the same counting-off experimental sections were then placed in new teams the semester. procedure that was employed at the beginning of administered to multiple choice examination was At the end of the semester a 60-item The UCRS and chapters assigned since midterm. all four classes covering only those second time. TMPA were also administered for the Results and Conclusions control combined experimental and combined T-tests were performed between the final examinations, and (13 quizzes, the mid-term and groups for all measurements Only one significant of the UCRS and the TMPA. the first and second administrations classes scored analyses. (See Table 3). The control difference appeared in all of these did the experimental sections. higher on a Calculation quiz than Educational Importance team it appears that being assigned to a new Under the conditions of this experiment, evaluating the course, effect on learning that subject matter, at midterm has little or no involves extra time and teammates. Since forming new teams or evaluating one's approach to STAD in is probably not an efficient record keeping for the instructor, it college classes. 5 7 References Reward. Paper read at STAD: Group Vs Individual Gnagey, W.J. (1988). College level meeting, Chicago. Midwestern Educational Research Association of Making STAD (1995). Process vs. Product Methods Gnagey, W.J. & Denoyer, J. W. Association Meeting, Midwestern Educational Research Team Awards. Paper read at Chicago. Student-constructed vs. (1993). A Comparison of Gnagey, W.J. & Navarro, N. Y Cycle. Paper read in Phase Two of the STAD Teacher- constructed Learning Activities meeting, Chicago. Educational Research Association at the Midwestern Criteria For STAD (1994). The Effects of Changing the Gnagey, W.J. & Navarro, N.Y. Association Meeting, Midwestern Educational Research Team Awards. Paper read at Chicago read at Midwestern (1991). STAD in College. Paper Gnagey, W.J. & Ostrowski, K. Chicago. Educational Research Association meeting, for STAD The Effects of Changing the Criteria Gnagey, W. J. & Ostrowski, K. (1992). read at Parallel Experiments. Paper Team Awards at Midterm: Two meeting, Chicago. Midwestern Educational Research Association Psychology Competitive Educational Sherman, L.W. (1986). Cooperative vs. 283-295. Teaching and Teacher Education, 2, Classrooms: A Comparative Study. Student Achievement? Cooperative Learning Increase Slavin, R.E. (1983). When Does Psychological Bulletin, 24_, 429-445. and Controversy. Cooperative Learning: Consensus Slavin, R.E. (1989). Research on Educational Leadership, 47, 52-54. New Jersey: Theory, Research and Practice. Slavin, R.E. (1990). Cooperative Learning: Prentice Hall. Educational Research on Cooperative Learning. Slavin, R.E. (1991). Synthesis of Leadership, 48 71-82. 8 6 Table 1 Rotated Factor Pattern for UCRS2 Factor 3 Factor 2 Item Factor 1 0.12490 0.69193 0.07836 A -0.30593 0.58192 0.36320 B -0.40395 0.58731 0.19670 C 0.06187 0.25255 D 0.77268 0.12786 0.71172 0.27006 E 0.79150 0.23254 0.10361 F -0.18749 0.45104 0.60411 G -0.00718 H 0.41617 0.69302 0.19525 0.76926 0.03367 I -0.32949 0.39456 0.64069 J -0.00112 -0.09374 0.73521 K 0.10489 0.07623 0.69043 L Percent of Variance Explained by each Factor Factor 3 Factor 2 Factor 1 11.11909 29.25415 31.31526 Table 2 Pre-Treatment T-Test Comparison of All Dependent Variables Yu t DF Treatment N* X ELD 3.4 13.8 Q1 83 C 0.68 167 0.42 3.3 14.0 Q1 86 E 2.7 14.8 Q2 83 C 0.13 165 2.6 1.52 15.4 Q2 84 E 29.7 5.9 Q3 C 82 0.60 166 0.52 6.2 Q3 29.2 86 E 2.5 13.4 Q4 C 81 0.30 164 1.03 2.6 12.9 Q4 85 E 2.3 13.6 Q5 C 79 0.57 0.56 2.7 161 13.9 Q5 84 E MID 41.7 7.4 C 81 0.27 163 1.11 7.3 40.4 MID 84 E UCRSI 7.9 45.8 C 75 0.56 145 0.59 UCRSI 6.8 46.5 72 E TMPAI 1.9 23.0 79 C 0.12 132 1.57 TMPAI 22.4 3.3 83 E 3.4 UCRSIL 1.4 83 C 0.27 167 1.11 1.7 UCRSIL 3.2 86 E * The variation in N is due to absences and withdrawals. 10

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.