ebook img

ERIC ED540962: Gathering Feedback for Teaching: Combining High-Quality Observations with Student Surveys and Achievement Gains. Policy and Practice Brief. MET Project PDF

5.4 MB·English
by  ERIC
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview ERIC ED540962: Gathering Feedback for Teaching: Combining High-Quality Observations with Student Surveys and Achievement Gains. Policy and Practice Brief. MET Project

MET Policy and project PracTicE BriEf Gathering Feedback for Teaching Combining High-Quality Observations with Student Surveys and Achievement Gains ABOUT THIS REPORT: This report is intended for policymakers and practitioners wanting to understand the implications of the Measures of Effective Teaching (MET) project’s interim analysis of classroom observations. Those wanting to explore all the technical aspects of the study and analysis also should read the companion research report, available at www.metproject.org. Together, these two documents on classroom observations represent the second pair of publications from the MET project. In December 2010, the project released its initial analysis of measures of student perceptions and student achievement in Learning about Teaching: Initial Findings from the Measures of Effective Teaching Project. Two more reports are planned for mid-2012: one on the implications of assigning weights to different measures; another using random assignment to study the extent to which student assignment may affect teacher effectiveness results. ABOUT THE MET PROJECT: The MET project is a research partnership of academics, teachers, and education organizations committed to investigating better ways to identify and develop effective teaching. Funding is provided by the Bill & Melinda Gates Foundation. Lead research and organizational partners include: ■ Mark Atkinson, Teachscape ■ Dan McCaffrey, RAND ■ Joan Auchter, National Board for Professional Teaching ■ Catherine McClellan, Educational Testing Service Standards ■ Denis Newman, Empirical Education ■ Nancy Caldwell, Westat ■ Roy Pea, Stanford University ■ Charlotte Danielson, The Danielson Group ■ Raymond Pecheone, Stanford University ■ Ron Ferguson, Harvard University ■ Geoffrey Phelps, Educational Testing Service ■ Drew Gitomer, Rutgers University ■ Robert Pianta, University of Virginia ■ Dan Goldhaber, University of Washington ■ Morgan Polikoff, University of Southern California ■ Pam Grossman, Stanford University ■ Rob Ramsdell, Cambridge Education ■ Heather Hill, Harvard University ■ Steve Raudenbush, University of Chicago ■ Eric Hirsch, New Teacher Center ■ Brian Rowan, University of Michigan ■ Sabrina Laine, American Institutes for Research ■ Doug Staiger, Dartmouth College ■ Michael Marder, University of Texas ■ John Winn, National Math and Science Initiative Acknowledgments: In addition to MET project partners who reviewed early drafts of the companion research report, we would like to thank the following external experts who read and provided written feedback on the document: Anthony Bryk, Andrew Ho, Bob Linn, Susan Moore Johnson, and Jonah Rockoff. The lead authors accept full responsibility for any remaining errors in the analysis.1 We want to express particular gratitude to the nearly 3,000 teachers who as MET project volunteers opened up their practice to help the project gain insights that can strengthen the teaching profession and improve outcomes for students. The lead authors of this report were Thomas J. Kane, Deputy Director of Research and Data at the Bill & Melinda Gates Foundation and Professor of Education and Economics at the Harvard Graduate School of Education, and Douglas O. Staiger, Professor of Economics at Dartmouth College. January 2012 Contents Guidance to Policymakers and Practitioners 2 Executive Summary 4 Defining Expectations for Teachers 7 Ensuring Accuracy of Observers 14 Ensuring Reliable Results 17 Determining Alignment with Outcomes 21 Three Key Take-Aways 29 A Not-Final Word 30 Endnotes 32 1 Gathering Feedback for Teaching Guidance to Policymakers and Practitioners Policymakers and practitioners at every level are intensely focused on improv- ing teaching and learning through better evaluation, feedback, and professional development. The Measures of Effective Teaching (MET) project is releasing these interim results because of that important work already under way in states and districts around the country. Although the project has much work still to do, the emerging findings have a number of important implications for the design of those systems. While classroom observations can play Therefore, we emphasize the following a central role in a teacher evaluation six minimum requirements for high- system by providing information for quality classroom observations:2 meaningful feedback, success hinges on 1. choose an observation instrument quality implementation. Good tools that that sets clear expectations. That are poorly implemented will have little means defining a set of teaching benefit. competencies and providing specific examples of the different perfor- mance levels on each. Many such instruments are already available and will be improving over time. Lengthy lists of vaguely described competencies are not sufficient. 2. require observers to demonstrate accuracy before they rate teacher practice. Teachers need to know that observers can apply an observation instrument accurately and fairly— before performing their first obser- vation. Good training is not enough. Observers should be expected to demonstrate their ability to generate accurate observations and should be recertified periodically. 22 MMeeaassuurreess ooff EEffffeeccttiivvee TTeeaacchhiinngg ((MMEETT)) PPrroojjeecctt 3. When high-stakes decisions are 5. combine observations with student 6. regularly verify that teachers being made, multiple observations achievement gains and student with stronger observation scores are necessary. For teachers feedback. The combination of also have stronger student facing high-stakes decisions, the classroom observations, student achievement gains on average. standard of reliability should be feedback, and student achievement Even a great observation instrument high. Our findings suggest that a carries three advantages over any can be implemented poorly. And any single observation cannot meet that measure by itself: (a) it increases the measure can become distorted in standard. Averaging scores over ability to predict if a teacher will have use. (This could be true for student multiple lessons can reduce the positive student outcomes in the feedback surveys as well.) Rather influence of an atypical lesson. future, (b) it improves reliability, and than rely on this study or any other (c) it provides diagnostic feedback as a guarantee of validity, school 4. Track system-level reliability by that a teacher can use to improve. systems should use their own data double scoring some teachers In the grades and subjects where to confirm that teachers with higher with impartial observers. At student achievement gains are not evaluation scores also have larger least a representative subset of measured, classroom observations student achievement gains, at least teachers should be observed by should be combined with student on average. impartial observers with no personal feedback surveys. relationship to the teachers. This is the only way to monitor overall system reliability and know whether efforts to ensure reliability are paying off. 3 Gathering Feedback for Teaching Executive Summary research has long been clear that teachers matter more to student learning than any other in-school factor. Improving the quality of teaching is critical to student success. Yet only recently have many states and districts begun to take seriously the importance of evaluating teacher performance and providing teachers with the feedback they need to improve their practice. The MET project is working with nearly surveys collecting confidential 3,000 teacher-volunteers in public student feedback; a new assessment schools across the country to improve of teachers’ pedagogical content teacher evaluation and feedback. MET knowledge; and different measures of project researchers are investigating student achievement. See figure 1. a number of alternative approaches In a previous paper, we reported that to identifying effective teaching: confidential student surveys about systematic classroom observations; students’ classroom experiences can provide reliable and meaningful feedback on teaching practice.3 In this report, we investigate the properties figure 1 of the following five instruments for classroom observation: MET Project ■ framework for Teaching (or ffT, Multiple Measures of Teaching developed by Charlotte Danielson of the Danielson Group), ■ classroom assessment Scoring System (or claSS, developed by S N T TIO EAC Robert Pianta, Karen La Paro, and NT PERCEP Teacher HER ASSES BVirrigdigneiat )H, amre at the University of STUDE OCBSlaEsRsVAroTIoOmNS SMENTS ■ PTeraoctohcinogl foobr slearnvgautiaognes a(orrt sP laTo, developed by Pam Grossman at Stanford University), Student Content ■ Mathematical Quality of instruction (or MQi, developed by Heather Hill of Harvard University), and STUDENT ASSESSMENTS 4 Measures of Effective Teaching (MET) Project ■ UTeach Teacher observation observation instruments alongside other teacher’s student achievement Protocol (or UToP, developed measures of teaching. gains with another group of by Michael Marder and Candace students as a measure that Walkington at the University of Key findings: combined observations with Texas-Austin). student feedback and achievement 1. all five instruments were positively gains on state tests. All the instruments establish a set associated with student achieve- of discrete competencies and then ment gains. 4. combining observation scores, describe observable indicators of differ- student feedback, and student The teachers who more effectively ent levels of performance. We studied achievement gains was better demonstrated the types of practices each instrument using two criteria: than graduate degrees or years of emphasized in the instruments had teaching experience at predicting 1. reliability. Reliability is the extent greater student achievement gains a teacher’s student achievement to which results reflect consistent than other teachers. gains with another group of aspects of a teacher’s practice and students on the state tests. not the idiosyncrasies of a particular 2. reliably characterizing a teacher’s observer, group of students, or practice required averaging scores Whether or not teachers had a lesson. over multiple observations. master’s degree or many years of experience was not nearly as power- In our study, the same teacher was 2. Validity. Validity is the extent to ful a predictor of a teacher’s student often rated differently depending which observation results are achievement gains on state tests on who did the observation and related to student outcomes. as was a combination of multiple which lesson was being observed. observations, student feedback, and If any of the instruments listed is to be The influence of an atypical lesson evidence of achievement gains with a helpful in practice, it will need to be and unusual observer judgment are different group of students. implementable at scale. To that end, reduced with multiple lessons and our analysis is based on 7,491 videos observers. 5. combining observation scores, of instruction by 1,333 teachers in student feedback, and student grades 4–8 from the following districts: 3. combining observation scores with achievement gains on state tests Charlotte-Mecklenburg, N.C.; Dallas; evidence of student achievement also was better than graduate Denver; Hillsborough Co., Fla.; New York gains on state tests and student degrees or years of teaching City; and Memphis.4 Teachers provided feedback improved predictive experience in identifying teachers video for four to eight lessons during the power and reliability. whose students performed well on 2009–10 school year. Some 900 trained Observations alone, even when other measures. raters took part in the subsequent scores from multiple observations lesson scoring. We believe this to be the Compared with master’s degrees were averaged together, were not as largest study ever to investigate multiple and years of experience, the reliable or predictive of a 5 Gathering Feedback for Teaching combined measure was better In addition, the combined measure The following pages discuss the able to indicate which teachers outperformed master’s and years instruments, scoring process, findings, had students with larger gains on of teaching experience in indicating and implications in greater detail. a test of conceptual understanding which teachers had students who Sections are organized around the in mathematics and a literacy test reported higher levels of effort and elements of the “Pathway to High- requiring short written responses. greater enjoyment in class. Quality Classroom Observations” in figure 2 below. figure 2 Pathway to High-Quality Classroom Observations as Part of a Multiple Measures System Define Determine Ensure Ensure EXPECTATIONS ALIGNMENT OBSERVER RELIABILITY FOR WITH ACCURACY OF RESULTS TEACHERS OUTCOMES 6 Measures of Effective Teaching (MET) Project Defining Expectations for Teachers Teachers, supervisors, and providers of professional development need a common vision of effective instruction to work toward. For this reason, the observation instruments in the MET project are not checklists focusing on easy-to-measure but trivial aspects of practice, such as whether or not lesson objectives are posted. Rather, for each defined competency the instruments require judgment about how well the observed practice aligns with different levels of performance. This is illustrated by the excerpts below from FFT. Indicators of Cognitive Challenge in “Use of Questioning and Discussion Techniques” from FFT UnSaTiSfacTory BaSic ProficiEnT diSTingUiShEd low cognitive Some questions reflect Variety of questions In addition to challenge, moderate cognitive challenge students and indicators in proficient predominantly challenge. advance high-level column, students recitation.5 thinking/discourse. initiate higher-order Ex: Teacher asks mix of questions. Ex: Teacher points to higher-order questions Ex: Most of teacher’s PowerPoint slide and and questions with single questions are open- Ex: A student asks of asks: “What does this correct answers. ended, as, “What might other students, “Does say?” have happened if the anyone have another idea colonies had not prevailed as to how we might figure in the American War for this out?” Independence?” 7 Gathering Feedback for Teaching Each instrument embodies a particular based on student understanding and Two patterns are clear in the study vision of effective instruction, reflected whether the teacher was teaching in sample. First, overall observed practice in the set of competencies on which his or her regular classroom. Similarly, is overwhelmingly in the mid-range of it chooses to focus attention and PLATO includes as many as 13 compe- performance as defined by the instru- how it defines proficiency in those tencies, but the PLATO Prime instru- ments. Second, scores are highest for competencies.6 The challenge for any ment that the MET project scored competencies related to creating an instrument developer is to identify a includes six. The version of MQI used, orderly environment and lowest for manageable set of competencies and MQI Lite, also included six competen- those associated with the most com- describe them with sufficient specificity cies, while the full version subdivides plex aspects of instruction. On FFT, to allow observers to score reliably. To those six into 24 elements that each for example, more than two-thirds of address this, the five instruments in the receives its own score. scores given for “managing student MET project take varied approaches to behavior,” “creating an environment of the total number of competencies and diagnosing Teacher respect and rapport,” and “engaging performance levels, as outlined in the Practice students in learning” are proficient or table on the following page. above. But the proficiency-and-above The different instruments paint a similar rate is just 44 percent for scores on Some of the instruments were stream- portrait of the nature of teaching in the “using assessment in instruction,” lined for the study, given the MET MET project classrooms. The potential 34 percent for “using questioning and project’s decision to score on a large that observations hold for providing discussion techniques,” and 30 percent scale using video. For example, the diagnostic information is evident in the for “communicating with students” (the version of FFT used in the study lacks distributions of scores of the MET proj- last competency requires clear pre- two competencies—or “components” ect volunteers. Observers used multiple sentation of content as well as cultur- as FFT calls them—found in the full measures to score the same 30 minutes ally and developmentally appropriate version of the instrument: Flexibility of instruction, and observers scored communication). Yet it’s these kinds of and Responsiveness & Use of Physical them in segments that varied accord- more complex teaching skills that will Space. We determined that these would ing to each instrument’s guidelines (15 be required for success on the new be difficult to score accurately without minutes for CLASS and 7.5 minutes for Common Core State Standards. conferring with the teacher about issues MQI, for example). figure 3, on pages such as how the lesson was planned The classroom observation’s greatest 10–11, shows the distribution of scores promise lies in its use as a developmen- given to lessons taught by MET project tal tool. To realize that promise, profes- volunteers for each competency on each sional development will need to be of the given instruments. individualized to meet teachers’ specific needs (just as content is being individu- alized to meet students’ needs in some schools today). The MET project has not yet investigated the impact of such new models of professional development, which are explicitly aligned with teach- ers’ evaluation results. However, there is encouraging evidence emerging from other research suggesting that such individualized feedback to teachers can lead to better outcomes for students.7 8 Measures of Effective Teaching (MET) Project

See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.