OEA's Teacher Evaluation & Support System Guidebook

54
Teacher Evaluation & Support System Guidebook OREGON EDUCATION ASSOCIATION Tools for Building an Effective, Collaboratively-Designed Evaluation System CENTER FOR GREAT PUBLIC SCHOOLS

description

Teacher Evaluation & Support System Guidebook - a tool for collaboratively developing a comprehensive teacher evaluation system in your school district.

Transcript of OEA's Teacher Evaluation & Support System Guidebook

Page 1: OEA's Teacher Evaluation & Support System Guidebook

Teacher Evaluation & Support System Guidebook

OREGON EDUCATION ASSOCIATION

Tools for Building an Effective, Collaboratively-Designed

Evaluation System

CENTER FOR GREAT PUBLIC SCHOOLS

Page 2: OEA's Teacher Evaluation & Support System Guidebook
Page 3: OEA's Teacher Evaluation & Support System Guidebook

3

Teacher Evaluation and Support System Guidebook

Gail RasmussenOEA President

Johanna VaanderingOEA Vice President

Richard SandersOEA Executive Director

Lindsey CappsDirector

Center for Great Public Schools

Teresa FerrerConsultant

Institute for Professional Skills & PracticeCenter for Great Public Schools

Erin WhitlockConsultant

Institute for Professional Skills & PracticeCenter for Great Public Schools

Oregon Education AssociationCenter for Great Public Schools

6900 SW Atlanta Street,Portland, OR 97223

1-800-858-5505www.oregoned.org

ACKNOWLEDGMENTSThe Center for Great Public Schools wishes to thank a number of individuals who have helped develop the foundation for this guidebook, which had its genesis in conversations among Oregon Education Association leaders, staff and other education stakeholders beginning in 2010.

Individuals we wish to acknowledge with gratitude are: OEA President Gail Rasmussen, OEA Vice President Hanna Vaandering, and (in alphabetical order), Sam Aley, Susan Anderson, Kristie Buckley, Daniel Burdis, Colin Cameron, Andrea Cooper, Tony Crawford, Linda Darling-Hammond, Dave Fiore, Nancy Golden, Anne Goff, Gary Humphries, Tom Husted, Henry Kim, Kathi Koenig, Kevin Mechlin, Colleen Mileham, Mark Molner, Jen Murray, Theresa Richards, Hilda Roselli, Richard Sanders, Eric Schutz, Monica Smith, Terrel Smith, Gwen Sullivan, Judy Svoboda, Joe Swinehart, Becca Uherbelau, Courtney Vanderstek, David Wilkinson, Colleen Works, and Robert Young.

Center of Great Public Schools staff Erin Whitlock, Teresa Ferrer, Center Director Lindsey Capps, and OEA’s Editor and Designer Meg Krugel are deserving of special acknowledgment for their research, writing, organizing and design of this Guidebook.

Page 4: OEA's Teacher Evaluation & Support System Guidebook

4

Executive Summary 6 Getting Started 9 a. 7 Principles for Success b. Union Leadership & Teacher Evaluation c. Collaboration as Key for the Development Process c. ESEA Waiver: Influencing Oregon’s Framework for TESS d. New Oregon Evaluation & Support Requirements e. Timeline, Process Checklist & Key Questions Tools 17 a. Design Process Checklist

Standards 25 a. Getting to Know the InTASC Standards b. The InTASC Teaching Standards at a Glance

Multiple Measures 27 a. Measures of Professional Practice b. Measures of Professional Responsibilities

Student Learning Goals 33 a. Student Learning and Growth as a Measure of Teaching Effectiveness b. Student Learning and Growth Goal-Setting Process

Using Measures of Student Learning 35 a. OEA’s Recommended Use of Student Data b. Measures of Student Learning c. Using Measures of Student Learning as a Measure of Teaching Effectiveness

Value Added Models 45 a. OEA’s Analysis of Usage of Value-Added Models b. Shared Principles for Use of Value-Added Models

Evaluation Procedures 49 a. Evaluation and Professional Growth Cycle b. Peer Assistance b. Teacher observation approaches

CD-ROM Resources 53 a. Alternative Compensation Resources b. Bargaining Resources c. Collaboration Resources d. Communication Resources e. Glossary f. InTASC Resources g. Multiple Standards Measures Resources h. NEA Resources

Table of Contents

I. II. III. IV. V. VI.

I.

II.

III.

IV.

V.

VI.

VII.

VII.

VIII.

VIII.

i. OEA Resources j. Research k. Rubric Resources l. Sample Frameworks m. Sample TESS n. Teacher Evaluation OARs o. References

Page 5: OEA's Teacher Evaluation & Support System Guidebook

5

How to Use this Guidebook

Oregon Education Association’s Teacher Evaluation and Support System Guidebook has been developed with the goal of providing local education association and school district leaders a comprehensive resource to uti-lize in working collaboratively together in designing a teacher evaluation and support system. Think of this

Guidebook as an owner’s manual for assembling and operating your new system. Whether your local association is just beginning this new journey or is well on the path toward a newly aligned teacher evaluation and support system, this Guidebook can prove useful in many ways.

Of course, before getting to the assembly and operation of your system, it is important to consider the purpose of that system. OEA and our educator members believe the purpose of any educator evaluation system is to support ongo-ing professional learning, growth and collaboration with the goal of continuous improvement in teaching and learn-ing. Therefore, this Guidebook is intended to empower educators to lead the way in creating evaluation and support systems that are holistic and transparent with clear focus on and alignment to high-quality, research-based standards of practice.

What you’ll find in this Guidebook:

• State and federal requirements for educator evaluations• Model core teaching standards • Measures and evidence of effective teaching• Research and resources to ensure evaluation systems are valid, reliable and fair• Tools to guide you in the design process, including models for collaboration and consensus-driven decision-making• Critical questions to keep in mind as you design, pilot and implement your new teacher evaluation and support

system.

5

Page 6: OEA's Teacher Evaluation & Support System Guidebook

6

The Oregon Education Associa-tion believes that every student in our public schools deserves competent, caring and effective

teachers. A sound teacher evaluation and support system is one tool to achieve this goal.

In the spring of 2010, the Oregon Edu-cation Association convened a special workgroup to discuss the role of evalu-ation in promoting teacher quality and student success (see enclosed CD-ROM). The group — representing local educa-tion association leaders and UniServ staff within OEA, together with school district leadership, higher education, and other stakeholders — developed a white paper to guide local and state conversations on teacher evaluation. A key finding of the workgroup was that teacher evaluation cannot exist in isolation and must be part of an aligned continuum of standards and supports from pre-service, hiring, assign-ment, induction, mentoring, professional growth and development, and teacher leadership.

The work group concluded that teacher evaluation must be tied to a strong, extended system of support for teachers providing impactful opportunities for on-going professional learning and collabora-tion across the career continuum — from pre-service, induction, and mentoring, to ongoing professional growth and devel-opment, and teacher leadership. Addi-tionally, the work group recognized the importance of an evaluation system being supported by the teaching and learning conditions at play in their classrooms, schools, and communities. Teacher evalu-ation is one part of a whole system that impacts student learning and all parts of the system must be improved for students to flourish.

For a teacher evaluation and support system to be effective it must be a ho-listic, valid and reliable tool. As will be demonstrated in more detail throughout this Guidebook, the proper design of an evaluation system is critical. If an evalu-ation and support system is to enhance student learning, then it must be focused on improving the knowledge, skills and

classroom practice of professional educa-tors based on established standards of teaching, customized collaboratively by the locally-based design teams. Focused standards alone will not elevate teaching and learning without an aligned system of professional learning and support that is robust, collaboratively designed and im-proved by teachers and districts working together toward that common goal.

An evaluation system must also be grounded within a professional culture where teachers – as colleagues, team members, mentors or master teachers – are continuously engaged in purpose-ful goal setting, professional learning and collaboration. As part of this profes-sional culture, teachers and other educa-tors should be engaged as collaborative partners in the design of all aspects of the evaluation system. The evaluation process itself requires training, resources, and time for obser-vation, analysis, and goal setting that is collaborative and individual to each education professional. This goes hand-in-hand with the evaluation system itself being strongly aligned to individual pro-fessional practice and ongoing, job-em-bedded professional development for all areas. Ultimately, this requires sustained leadership and commitment – financial, social and political – from the school district, community, the state and elected leaders at all levels.

As school district teams of teachers, prin-cipals, other district administrators and the exclusive bargaining representatives work together to collaboratively develop evaluation systems around these core propositions, the result will not just be better evaluation systems. These collab-orative processes and resulting evaluation and support systems can create better professional learning environments where teaching and learning – and our students – thrive.

REDESIGN OF TEACHER AND ADMINISTRATOR EVALUATION SYSTEMS REQUIRED BY LAW

In the summer of 2011, Oregon Governor John Kitzhaber signed into law Senate

Bill 290, legislation supported by the Oregon Education Association and passed with bipartisan majorities in both cham-bers of the Oregon Legislature. Almost a full year later in July 2012, the US Depart-ment of Education approved Oregon’s re-quest for a waiver to the federal Elemen-tary and Secondary Education Act (ESEA) and No Child Left Behind law.

SB 290 and the provisions of Oregon’s ESEA waiver related to educator ef-fectiveness are predicated upon long-established evidence demonstrating that a teacher is the most important in-school factor impacting a student’s success. SB 290 and the Oregon waiver together establish new state policies aimed at im-proving the quality of both teachers and school leaders through new requirements and standards for evaluation and support systems. Both SB 290 and Oregon’s waiv-er will guide the work of locally-based collaborative design teams as they design their evaluation and support systems.

SENATE BILL 290: Collaborative Design of Standards-Based Evaluation Systems with Multiple Measures of Performance

Senate Bill 290 directed the Oregon State Board of Education to adopt standards for teacher and administrator evalua-tions. In December 2011, the State Board of Education adopted model core teach-ing standards developed by the Interstate Teacher Assessment and Support Con-sortium, known as the InTASC Standards. The State Board adopted standards established by the Interstate School Lead-ers Licensure Consortium, or the ISLLC Standards, for all administrator evalua-tions.

The standards adopted for both teach-ers and administrators are high-quality, research-based standards that reflect what an educator should know and be able to do.

In addition to setting forth statewide standards of practice for teachers and administrators, Oregon’s Senate Bill 290 has three key requirements for school districts and education service districts:

Executive Summary

Page 7: OEA's Teacher Evaluation & Support System Guidebook

7

Executive Summary

• Teacher evaluation systems must be designed collaboratively with teachers and their exclusive bargaining repre-sentative (local education association representing teachers in the district). [See Getting Started section for re-sources to assist in this process].

• By July 2013, every school district and education service district in Oregon must align their teacher evaluation systems with model core teaching standards adopted by the State Board of Education [See Standards section for in-depth, detailed information on the State adopted InTASC Standards].

• Aligning their teacher evaluation systems to state standards, and the application of standards across the evaluation and support system, school districts must create evaluation systems that take into consideration multiple measures of teaching effectiveness, and establish a formative growth process for each teacher that supports profes-sional learning and collaboration with other teachers. [See Multiple Measures section for a discussion of measures to consider including in your district’s evaluation system; and how to inte-grate measures within your evaluation system].

OREGON’S ESEA WAIVER: Comprehensive Framework for Evalua-tion Systems Integrating Standards and Multiple Measures

In providing Oregon the opportunity to obtain a waiver from the onerous provi-

sions and punitive sanctions imposed un-der the federal No Child Left Behind law, the US Department of Education required that the State and local school districts “… commit to develop, adopt, pilot, and im-plement, with the involvement of teach-ers and principals, teacher and principal evaluation and support systems.”

In receiving approval of its waiver appli-cation from the US Department of Educa-tion in July 2012, Oregon committed itself to a set of basic guidelines for teacher and principal evaluation and support systems. These guidelines are reflected in the Framework for Teacher and Administra-tor Evaluation and Support Systems ad-opted by the State Board of Education and provided to all Oregon school districts in June 2012.

Oregon’s ESEA waiver expands on Senate Bill 290 with four additional require-ments:• School districts must comply with state

framework for teacher and administra-tor evaluation and support systems. [See Framework for Teacher & Administra-tor Evaluation & Support section].

• Establishes 3-specific categories of mea-sures to be included in teacher evalua-tion and support systems: Professional Practice, Professional Responsibilities and Student Learning and Growth. [See Multiple Measures section].

• Incorporate student growth as a “sig-nificant factor” in individual teacher evaluations, requiring school districts to provide teachers the opportunity for

individual goal setting around student learning. [See Measures of Student Learning section for strategies to meet this requirement in ways that are mean-ingful and educationally-relevant based on subject area and assignment.] ODE will be doing a pilot year in 2012-13 to determine what “significant” will mean in the course of teacher and adminis-trator evaluations; more information about this factor will be forthcoming in the spring of 2013.

• Four-levels of proficiency in assigning a summative rating of performance to each individual teacher.

Under the waiver, Oregon committed it-self to a pilot year for 2012-2013 in which 50 sites will be participating. Schools involved in the federal School Improve-ment Grant (SIG) program, the Teacher Incentive Fund (TIF), or the state School District Collaboration Grant Program will be potential candidates, and other sites may be allowed to apply. These pilots will integrate the student learning goals [see the Using Measures of Student Learn-ing section] into their evaluation systems and determine what “significant” means per the ESEA Waiver criteria. Further-more, school districts not participating in the ODE pilot year, have until the 2013-2014 school year to pilot their evaluation system in compliance with SB 290 and the waiver. [See additional ESEA waiver re-quirements and timelines in the Getting Started section].

Teacher Evaluation = Three “Buckets” of Evidence

TEACHING & LEARNING CONDITIONS

PROFESSIONAL PRACTICE

ObservationsArtifacts

PROFESSIONAL RESPONSIBILITIES

Self Assessment Leadership Roles

Prof. DevelopmentFamily Engagement

Growth Plans

STUDENT GROWTH & LEARNING

Student learning goals based on

multiple measures

Page 8: OEA's Teacher Evaluation & Support System Guidebook

8

PROFESSIONAL PRACTICE

PROFESSIONAL RESPONSIBILITIES

STUDENT LEARNING

Triangulated Standards-Based

Evaluation Framework

↕ ↕↕

Validates Judgments about Practice

Validates Judgments about Practitioner

Oregon’s Framework for Teacher and Administra-tor Evaluation and Support Systems established under the ESEA waiver is grounded in the concept

triangulation. Triangulation is used in many complex “hu-man” fields where you cannot easily or accurately assign a quantitative number or value. The logic behind triangula-tion is that no single measure reveals sufficient or reliable information. Multiple data points on the other hand pro-vide richer and more accurate results. Conclusions based on one measure are vulnerable to erroneous conclusions, while multiple pieces of evidence using different types of data are more likely to result in valid conclusions.

The use of triangulation in the evaluation of teachers is a valuable approach because it accounts for the complex landscape of teaching and learning. It can be used as a model for evaluating what teachers do in their practice and profession because it allows for “performance” to be viewed from more than one angle. Under triangulation, multiple measures are used to gather information about both the practice and the practitioner, including observa-tions, artifacts, teacher reflection on professional develop-ment or goals, and student work, among other evidence. In this way, triangulation is a more concrete method for understanding teaching effectiveness.

So how does triangulation work in practice? Triangu-lation requires a method or measure of teaching ef-fectiveness to be validated by others, leading to greater confidence in the overall feedback an educator receives. Different methods and measures are designed to vali-date one another and support the same conclusion. This process will lead to teachers and their evaluators having greater confidence in evaluation system while reducing subjectivity.

When agreement is reached between measures from the three different categories required under Oregon’s Framework, there is greater confidence that the original assessment was accurate. If the data sources differ, that triggers a need for the evaluator to assess the training system in place to train evaluators on the measure, the measure itself, and/or the educator again more deeply to figure out why results varied. Triangulation also ad-dresses the limitations of each source of information. Judgments about practice are limited by subjectivity and bias, while judgments based on student performance are limited by technical considerations.

Adapted from Massachusetts Teacher Association

Inside Oregon’s Framework: Triangulation of Multiple Measures to Support the Whole Teacher

Executive Summary

Page 9: OEA's Teacher Evaluation & Support System Guidebook

9

Getting Started

OEA believes that seven key princi-ples must serve as the foundation for developing or reforming any

Teacher Evaluation and Support System (TESS):

1. Safe and open collaboration is necessary. When assessment of teacher practices is transparent and openly collaborative, teachers can build profes-sional communities and learn from one another. This process can occur only in non-threatening environments of forma-tive assessment and growth.

2. Measures of teacher performance are most helpful and meaningful when they are based on multiple valid ratings and clear teaching standards. Teachers need clear and actionable feedback based on standards for teaching and student learning that are comprehensive and transparent, and on criterion-referenced assessments of teacher practice. Feedback is most useful as part of a comprehensive teacher development system. Summa-

tive evaluations of teachers should use uniform criteria for effectiveness that are relevant for all teachers.

3. Integrated systems must link evaluation procedures with curricular standards, professional development activities, targeted support, and person-nel decisions.

4. Validated evaluation measures are essential. Measures of teaching effective-ness need to be based on widely accepted standards of teaching that capture a range of teaching behaviors and use multiple valid evaluation methods.

5. Teachers’ input in determining per-formance and learning outcomes should be part of the system. Although standards for teaching practice and student learning are essential, each teacher should have an opportunity to help define a set of prac-tices and student learning objectives to be assessed. Teacher input can provide vital learning goals for the unique, contextu-

alized circumstances of each particular classroom.

6. Teacher Evaluation and Support Systems (TESS) need to be co-created or designed with teachers working through the local association. This may be the most important principle of all. Ideals and visions need to be balanced with local context, and political and financial reality. There is no one-size-fits-all solution at a state level. OEA will work with locals to craft local solutions based on the prin-ciples outlined here.

7. Teacher Evaluation and Support Systems are found to be more effective when:• they ensure that evaluators are well-

trained• evaluation and feedback are frequent• mentoring and coaching are available• processes are in place to support due

process and timely decision making by an appropriate body.

I.

7 Principles for Success

Student learning is at the center of everything a teach-er does. OEA believes that effective teaching engages all students in the learning process; focuses on inter-

actions and activities between teachers and students, and students with their peers; involves collaboration among teachers; centers on a continuous professional learning cycle where planning, practice, implementation, reflec-tion, analysis, and modification of practice occur; and leads to growth in student knowledge, skills, and well-being. Evaluation of this process needs to be flexible, robust, and based on the value of continuous growth and improvement in the profession. OEA’s resolutions exemplify this value for our association, making advocacy for the active assess-ment of all teachers through regular and comprehensive

evaluation procedures a goal to which we strive (OEA Resolution V.14). Our members believe this needs to be implemented in a professional environment which respects the diversity of our students and workforce, and built from the perspective that evaluations should be fair and objective for all school employees, and should be developed by, and acceptable to, the association and the governing board in compliance with state law (OEA Resolution V.22). Additionally, OEA’s Core Values of lifelong learning, collaboration, respect for diversity, integrity, and professionalism insist that the evaluation systems we co-create exemplify these values in every aspect.

UNION LEADERSHIP & TEACHER EVALUATION

“OEA’s Core Values of lifelong learning, collaboration, respect for diversity, integrity, and professionalism insist that the evaluation systems

we co-create exemplify these values in every aspect.”9

Page 10: OEA's Teacher Evaluation & Support System Guidebook

10

Collaboration is the standard for the design and implementa-tion process. This means there needs to be some standard-

ized methods of engaging in this work that create equal access, opportunity, and voice to all participants. OEA has some key rec-ommendations in establishing a collaborative process to design and implement your TESS (see CD-ROM for activities and ideas suggested below):

1. Establish a process for coming to a consensus-driven decision.

2. Develop group norms to guide and enhance partici-pants’ behaviors throughout the process.

3. Establish a procedure for disagreements (how to make a consensus-decision when there is NOT a consensus, such as bringing in an outside facilitator)

4. Start with lower-stakes activities when beginning with a consensus-based process; working collaboratively takes trust, and trust takes time to build over a continuous cycle where each “side” takes risks, is asked to follow-through on certain important tasks/commitments, then follows through with their responsibilities and is the recipient of the “other side’s” trust for having done so:a. Collaborate on a group definition of effective teaching.

This can be done separate from Oregon’s definition of effective teaching, which is effectively the InTASC Stan-dards (see below for more information), and can work to enhance and inform these standards so that districts can customize what they also find important for assessing teaching as well as customize the application of standards across the TESS.

b. Collaborate on a group set of goals for the teacher evalua-tion and support systems.

c. Come to consensus on a common vocabulary. There are many terms in developing a TESS that are misunderstood or misused (student achievement vs. student growth is one resounding example), so looking at the Glossary of Terms in the InTASC Standards is helpful. Making sure everyone is on the same page for what important terms mean is one way to build up trust and to prevent future misunderstandings.

d. Come to consensus on a common understanding of the InTASC Standards, Oregon’s new Model Core Teaching Standards to which all teacher evaluation systems must be aligned by July 1, 2013.

Collaboration as Key for the Development Process

Getting StartedI.

EVALUATION

Student Achievement

ProfessionalDevelopment

Collaborative Culture

Page 11: OEA's Teacher Evaluation & Support System Guidebook

11

Getting Started

ESEA WAIVER FRAMEWORKSThe following is a summary of the ESEA Waiver requirements that must be a part of your Teacher Evaluation and Support System (TESS):

1. Require the TESS be used for con-tinual improvement of instruction

2. Require four performance level rat-ings of effectiveness (see chart below)

3. Require the TESS have multiple, valid measures in determining perfor-mance levelsa. This will include a rubric or perfor-

mance indicator to clarify perfor-mance expectations for each of the ten InTASC Standards. The rubric must have the required four perfor-mance level ratings of effectiveness.i. If other frameworks are used, such

as the Danielson frameworks, a “crosswalk” must be made to clarify which of the InTASC Stan-dards are covered by each domain/criteria.

b. This will include choosing at least one measure from each category of measures:i. Professional practiceii. Professional responsibilities and iii. Student learning

c. This will also include a robust set of measures of student learning for all students as a “significant factor” in teachers and administrators’ evalua-tions. The use of student data will be discussed in more detail in this OEA Guidebook.

i. Oregon Department of Education (ODE) will be conducting pilots in 2012-2013 to determine what “significant” means.

4. Require teachers to be evaluated on a regular basis

a. Probationary teachers: annuallyb. Contract teachers: 2-year cycle

5. Require clear, timely, and useful feedback to guide professional develop-menta. Establish a formative growth process

for teachers that supports profes-sional learning and collaboration with other teachers and administra-tors

b. Use evaluation methods and profes-sional development, support and other activities that are based on curricular standards and that are targeted to the needs of each teacher

c. Align professional development op-portunities with educator self-reflec-tion, assessment, and goal-setting.

d. The focus of local evaluation and support systems is to help educators improve their practice to improve student learning. Collaborative teams should determine what kind of support a teacher or administrator can expect if they are not proficient on all standards.

e. Professional learning should be aligned to the TESS. The Oregon Framework suggests professional learning may be guided by the Learn-ing Forward standards (see CD-ROM); job-embedded, collaborative,

and customized to individual educa-tor needs. The standards assert that professional learning that increases educator effectiveness and results for all students must include:1. Learning Communities: occurs

within learning communities committed to continuous improve-ment, collective responsibility, and goal alignment.

2. Leadership: requires skillful lead-ers who develop capacity, advo-cate, and create support systems for professional learning.

3. Resources: requires prioritiz-ing, monitoring, and coordinating resources for educator learning.

4. Data: uses a variety of sources and types of student, educator, and system data to plan, assess, and evaluate professional learning.

5. Learning Designs: integrates theories, research, and models of human learning to achieve its intended outcomes.

6. Implementation: applies re-search on change and sustains support for implementation of professional learning for long term change.

7. Outcomes: aligns its outcomes with educator performance and student curriculum standards.

6. Will be used to inform person-nel decisions. School districts must describe in policy how their TESS is used to inform personnel decisions (e.g., contract status and renewal, plans of assistance, placement, assignment, career advancement, etc.).

I.

ESEA Waiver: Influencing Oregon’s Framework for Teacher Evaluation & Support Systems

Performance Levels Definitions of Performance as Applied to Standards of Professional Practice1 Does not meet this standard

2 Making sufficient progress toward meeting this standard

3 Consistently meets expectations for good performance under this standard

4 Consistently exceeds expectations for good performance under this standard

Page 12: OEA's Teacher Evaluation & Support System Guidebook

12

Evaluation systems used to determine e�ectiveness of educators & in making human resources decisions *School districts must develop policy to say how the evaluation & support system is used to inform personnel decisions

Include job descriptions and related performance standards

Based on written criteria(including performance goals) (ORS 342.850 (2))

Evaluation systems used to improve professional development of educators

Include multiple measures of educatore�ectiveness

Based on at least two observations in combination with other information(ORS 342.850 (1))

Evaluation systems used for continual improvement of instruction

Include student progress based on multiple measures (including student, school & performance measures)

Includes pre & post evaluation interviews(ORS 342.850 (2))

Evaluation systems designed jointly by district, teachers and exclusive bargaining representative

Built on research-based practices

If needed: includes written plans of assistance(ORS 342.850 (2))

Evaluations systems aligned

with InTASC Standards and customized application of standards throughout evaluation and support system.

Separately developed for teachers & administrators

Evaluation reports maintained in district personnel files, but only after reasonable notice to teacher(ORS 342.850 (4), (5))

*Evaluation systems use four levels of performance

Customized for each district

*Evaluation systems use multiple & valid measures in determining performance levels of educators, including, as a significant factor, student performance (defined locally and through ODE pilot process)

Allow for individual di�erences in assignment

Educator may attach their own written statement to any evaluation document in file(ORS 342.850 (6))

*Student performance will be evaluated via multiple measures embedded within collaboratively-designed student learning goals.

Used to refine support & professional growth system based on needs of individuals, schools & district*Teacher evaluation & support systems must align to professional development Oregon Framework suggests Professional development must be guided by NSDC/Learning Forward Standards.

Establish formative growth process that supports professional growth &collaborative learning

School boards will adopt policy specifying which school o�cials have file access (ORS 342.850 (9))

Use methods based on curricular

standards targeting the needs of individual educators

Check your Collective Bargaining Agreement for other requirements

New Oregon Evaluation & Support Requirements

*Denotes compliance policies that are part of Oregon’s ESEA Waiver and are additional requirements above our newly revised Evaluation statutory requirements.

The following chart can serve as a checklist for your Teacher Evaluation and Support System (TESS) design teams. The chart presents compliance pieces from new state-level policy (SB 290), the ESEA Waiver, and already established Oregon Statutes and Oregon Administrative Rules, which

your team can use to ensure your TESS is a collaboratively designed system that will measure up to all the requirements.

New State Law & Policy

Collaboratively Designed System

Pre-Existing State Law Bargaining Requirements

New State Law & Policy

Collaboratively Designed System

Pre-Existing State Law Bargaining Requirements

Page 13: OEA's Teacher Evaluation & Support System Guidebook

13

Evaluation systems used to determine e�ectiveness of educators & in making human resources decisions *School districts must develop policy to say how the evaluation & support system is used to inform personnel decisions

Include job descriptions and related performance standards

Based on written criteria(including performance goals) (ORS 342.850 (2))

Evaluation systems used to improve professional development of educators

Include multiple measures of educatore�ectiveness

Based on at least two observations in combination with other information(ORS 342.850 (1))

Evaluation systems used for continual improvement of instruction

Include student progress based on multiple measures (including student, school & performance measures)

Includes pre & post evaluation interviews(ORS 342.850 (2))

Evaluation systems designed jointly by district, teachers and exclusive bargaining representative

Built on research-based practices

If needed: includes written plans of assistance(ORS 342.850 (2))

Evaluations systems aligned

with InTASC Standards and customized application of standards throughout evaluation and support system.

Separately developed for teachers & administrators

Evaluation reports maintained in district personnel files, but only after reasonable notice to teacher(ORS 342.850 (4), (5))

*Evaluation systems use four levels of performance

Customized for each district

*Evaluation systems use multiple & valid measures in determining performance levels of educators, including, as a significant factor, student performance (defined locally and through ODE pilot process)

Allow for individual di�erences in assignment

Educator may attach their own written statement to any evaluation document in file(ORS 342.850 (6))

*Student performance will be evaluated via multiple measures embedded within collaboratively-designed student learning goals.

Used to refine support & professional growth system based on needs of individuals, schools & district*Teacher evaluation & support systems must align to professional development Oregon Framework suggests Professional development must be guided by NSDC/Learning Forward Standards.

Establish formative growth process that supports professional growth &collaborative learning

School boards will adopt policy specifying which school o�cials have file access (ORS 342.850 (9))

Use methods based on curricular

standards targeting the needs of individual educators

Check your Collective Bargaining Agreement for other requirements

New Oregon Evaluation & Support Requirements

*Denotes compliance policies that are part of Oregon’s ESEA Waiver and are additional requirements above our newly revised Evaluation statutory requirements.

The following chart can serve as a checklist for your Teacher Evaluation and Support System (TESS) design teams. The chart presents compliance pieces from new state-level policy (SB 290), the ESEA Waiver, and already established Oregon Statutes and Oregon Administrative Rules, which

your team can use to ensure your TESS is a collaboratively designed system that will measure up to all the requirements.

New State Law & Policy

Collaboratively Designed System

Pre-Existing State Law Bargaining Requirements

New State Law & Policy

Collaboratively Designed System

Pre-Existing State Law Bargaining Requirements

Getting Started I.

Process Checklist & Key Questions for Collaboratively Developing a Teacher Evaluation and Support System (TESS)

SHORT OUTLINE (see next page for full process)

I. BEFORE DESIGN PROCESS a. Designate members of your bargaining unit to be on Teacher Evaluation Design team i. Have your local President and Bargaining Team meet with the Superintendent ii. Consult OEA’s Center for Great Public Schools for any supports/trainings you may need b. Design a collaborative process for the meetings c. Schedule meetings on a timeline to finish process by July 1, 2013 d. Design a communications plan

II. DESIGN PROCESS 1. The Teacher Evaluation and Support System (TESS) framework 2. The rubric used for this framework 3. Multiple valid measures of teaching effectiveness (which includes multiple valid measures of student learning, professional practice, and professional responsibilities) 4. Professional Development and Growth System (PDGS) that is informed by teacher evaluation 5. Reworking or revising of current job descriptions and performance standards as necessary 6. Training of both administrators and teachers in the new system 7. Pilot & Implementation or roll-out plan 8. Evaluation and continuous improvement plan

School Year Activities2011-12 • State Board adopted state framework.

2012-13

• ODE pilots framework at selected sites (SB 252, SIG, others) to determine what “significant” use of student learning data in teacher and administrator evaluation means.• All districts develop local evaluation and support systems consistent with state guidelines/ framework.

By July 1, 2013 • All districts submit revised evaluation and support systems, implementation plans, and training plans.

2013-14 • All districts pilot implementation of local evaluation and support systems.

2014-15 • All districts fully implement local evaluation and support systems.

By July 1, 2015 • All districts present local evaluation and support systems to a Regional Peer Review Panel.

Oregon Framework Timeline for Implementation

Page 14: OEA's Teacher Evaluation & Support System Guidebook

14

Pre-Design Process Checklist & Key Questions

In working through each design section, the Teacher Evaluation Design (TED) Team can use the following list of process elements and key questions to guide

and focus their discussions and work. The TED team may determine that more comprehensive external feedback from individual stakeholder groups is necessary to con-tinue. Some of these questions might also serve as useful starters for a membership survey, so feel free to use them in any format you see fit (See also Laura Goe’s “Questions to Ask About Measures and Models” on the CD-ROM):

BEFORE DESIGN PROCESS Access OEA’s Teacher Evaluation Guidelines and

supporting documents.

Executive Committee designate one or more people to create a Teacher Evaluation Planning Committee (TEPC) to oversee this work and adds TEPC report to every Executive Committee agenda for the next two years at least (at least one member of the Executive Committee must be on TEPC)

Add members to committee so that it is made up of representatives from:

1. Different grade level buildings

2. Different specialty/content areas

3. Different demographics (experience, race/ethnicity, age, etc.)

4. One person+ on TEPC must have bargaining experi-ence or be a bargaining team member

5. Include Special Education and English Language Development personnel in designing, implementing, and monitoring evaluation models

Teacher Evaluation Planning Committee assign at

least three members in the committee to form a sub-committee in charge of internal communication (see OEA guidelines)

1. Clarify expectations in terms of stakeholder groups’ purposes and authority in decisions

2. Define stakeholders’ roles and responsibilities in a way that capitalizes on their expertise

3. Create a two-way collaborative communication plan for all external (outside of group) communication (to members, parents, students, community, etc.)

4. Create a two-way communication system to distrib-ute and gather information from membership

5. Consider the content, the target audience, and the mode and timing of these communications, as well as what will be done with the responses.

Teacher Evaluation Planning Committee looks over OEA materials, attends regional training or invites Center for Great Public Schools staff to give overview to committee

Teacher Evaluation Planning Committee initiates contact with district to set up *collaborative design process for developing the new Teacher Evaluation and Support System (TESS)*

Teacher Evaluation Planning Committee holds initial meeting with their Bargaining Team to give overview of collaborative design process and discuss placehold-er language options and/or how this process will be represented (via an MOU, contract language, separate Teacher Evaluation and Support System handbook, etc.)

1. Decide how the bargaining team and TEPC will be communicating with one another throughout the process.

Teacher Evaluation Planning Committee (or repre-sentatives of TEPC) and district team form the new Teacher Evaluation Design Team (TED Team) so that teachers are at least equally represented.

Teacher Evaluation Design Team sets up regular meetings at least monthly that are contractually sup-ported

Teacher Evaluation Design Team shares resources

and spends time clarifying the following:

1. Group norms

¢ How do you revisit/reinforce them

I. Getting Started

14

Page 15: OEA's Teacher Evaluation & Support System Guidebook

15

Getting StartedI.

2. What the collaborative process will look like

¢ Decide how you reach agreements

¢ What to do when you agreements are not met

¢ Decide how you maintain balance between parties

3. Sharing of responsibilities

¢ Determine what roles will need to be filled (facilitator, note-taker, public scribe, time-keeper, “temperature-reader,” group norms sentinel, etc.) and record each roles job description

4. How minutes are taken and shared with entire district

5. How entire district is allowed to contribute to the work of the team (increases buy-in)

6. End game goals and outcomes

¢ Define the purpose of the Teacher Evaluation and Support System.

¢ Make sure goals are stated in measurable terms which are explicitly, well-defined, and easily under stood by all stakeholder groups.

¢ Make sure the Teacher Evaluation and Support Sys-tem goals are aligned to meet the requirements of state policy and research-based best practices.

¢ Determine resources necessary to complete Teacher Evaluation and Support System design process

¢ See “Evaluation and Continuous Improvement Plan” notes below

7. *Timeline for getting system design completed by July 1, 2013*

¢ Build in communication/feedback loops with entire district into timeline

8. Define stakeholder groups9. Beginning Collaborative Decisions

¢ Collaborate on a group definition of effective teaching

¢ Come to consensus on a common vocabulary

¢ Come to consensus on a common understanding of the InTASC Standards

* Asterisks represent legal requirements

15

Page 16: OEA's Teacher Evaluation & Support System Guidebook

1616

Page 17: OEA's Teacher Evaluation & Support System Guidebook

17

Teacher Evaluation Design Team divides the work of the de-sign into the following eight sections and decides how to move through each section and keep the comprehensive design aligned and cohesive:

1. The Teacher Evaluation and Support System (TESS) framework *How it will be used (formative and summative)*

*Determine the frequency of evaluation (align to state policy) and the frequency of each measure used*

¢ How many observations using the framework per year (different for various career levels?)

¢ How it is informed by written performance goals

¢ How it is informed by pre- and post-observation inter-views

*How it will be aligned to and improve Professional Development*

*How it is included within a Professional Growth System*

*How it is aligned to InTASC Standards and the applica-tion of the Standards across the entire evaluation and support system*

*How it will have multiple measures*

How it will be revised to apply to different specialties/assignments

Consider determining a base-line level of performance prior to making decisions regarding teacher proficiency levels

Consider the technical defensibility of each measure and the Teacher Evaluation and Support System as a whole to make personnel decisions (or compensation decisions if you are in a school district with TIF/SIG funds)

Compare and contrast your new Teacher Evaluation and Support System framework with your current TESS to see what (if anything) is worth keeping and what needs to be tweaked, modified, or thrown away altogether

Identify a common framework for effectiveness and in-clude differentiated criteria where applicable, especially when considering teachers’ various assignments and

specialties (such as Special Education and English Lan-guage Development personnel)

Integrate evidence-based practices for teachers working with students who are English language learners and stu-dents with disabilities

Ensure that evaluation framework can identify and pro-vide targeted Professional Development

Ensure that TESS Framework meets the Oregon Frame-work for Teacher and Administrator Evaluation and Sup-port Systems (see CD-ROM for Framework).

◗ KEY QUESTIONS TO ASK:

Where is the money going to come to support:¢ These collaborative conversations?«Release time

¢ The training of all teachers/certified staff and admin-istrators/evaluators?« Time? Pay?

¢ The sustainability of an enhanced system?

If evaluation of teachers is to measure the ability of the teacher to teach, what is our shared definition of effec-tive teaching?¢ This definition is important to use as a reflective tool

to make sure your Teacher Evaluation and Support System aligns to this definition

What are our shared agreements about how a Teacher Evaluation and Support System should work?

Does the process allow for the cooperative development and assessment of an evaluation system?

Is there a system of checks and balances for the process?

What does our current Teacher Evaluation and Support System look like?¢ What are its strengths? What works?¢ What are its weaknesses? What doesn’t work?¢ Is it aligned to any professional development sys-

tem?¢ How often is it assessed by an inclusive group?¢ Is it manageable? Clear to everyone? Effective in

growth and improvement of practice?¢ How does it apply across levels of experience/con-

tent area/specialists?« Is it differentiated across levels of experience/

content area/specialists?¢ *Are there multiple valid measures of teaching effec-

tiveness?*

DESIGN PROCESS CHECKLIST

* Asterisks represent legal requirements

II.Tools

17

Page 18: OEA's Teacher Evaluation & Support System Guidebook

« *Does that include multiple valid measures of stu-dent performance? *Must have at least one mea-sure of student performance*

« Are there established teacher evaluation tools you can modify and adapt to these purposes?

¢ Does each tool align to InTASC Standards, and does each customized application of the Standards across the entire evaluation system reflect the integrity of In-TASC, or are there crosswalks available?« How will the district and/or Association evaluate

the validity, reliability, comparability, relevancy, meaningfulness of the measures being used? Some questions you might ask yourselves are:

¢ Does the evaluation system achieve the purposes for which it was designed?

¢ How well does the evaluation system support effec-tive teaching for all students?

¢ Are the measures valid in terms of measuring the agreed-upon definition of effective teaching?

¢ Do the measures meet high standards of reliability in every school and for every teacher?« What current bargaining language do you have re-

garding your evaluation system?¢ What are its strengths?¢ What are its weaknesses?¢ How will disputes of findings be processed?¢ See bargaining section of OEA’s Teacher Evaluation

and Support Systems Guidebook

2. The rubric used for this framework *Required four levels of performance*

Define what the triggers are for moving into and out of plans of assistance (POA)

Define trigger points for action (professional growth, pro-fessional development, etc.) – when action will be trig-gered, what kind-of action, for how long, etc.

Determine how evaluation results will be shared with teachers and when they will be given notice of the next steps toward professional growth or termination

Determine how much time and assistance will be provided for a teacher to demonstrate improvement before termina-tion is considered (assure that POA language and proce-dures align to this)

*Rubric must be clearly delineated for all ten InTASC Stan-dards at all four performance levels, or a crosswalk must

be made to outline where each of the InTASC Standards is addressed.*

◗ KEY QUESTIONS TO ASK:

Does the process provide a system of support to assist members to move from one performance level to anoth-er? Have the expectations been clearly articulated for each performance level and the triggers for progressing along the continuum?

Does the process allow for reasonable, attainable iden-tification of ways to correct significant discrepancies/move out of plans of assistance that is also aligned with supports?

What would it look like for your teacher evaluation system to differentiate across specialty areas and assignments?¢ What types of specialists are a part of your bargain-

ing unit?¢ How do their day-to-day jobs differ from that of a

classroom teacher?« Clientele?« In planning and preparation?« The environment?« Delivery of service?« Professional Responsibilities?

¢ What assurances need to be considered or in place to assure equity of the evaluation system across spe-cialty areas and assignments?« What would it look like for your teacher evaluation

system to differentiate across years of experience?¢ What supports do newer teachers need?¢ How could you graduate your system to meet the

needs of all levels of experience?

3. Multiple valid measures of teaching effectiveness (which includes multiple valid measures of student learning, professional practice, and professional re-sponsibilities) *At least one measure identified for each category of teach-

ing effectiveness: professional practice, professional re-sponsibilities, and student learning (which will be the Stu-dent Learning Goal as outlined on page 33).*

Measures are research-based for the purpose of individual teacher evaluation

Measures are relevant; they measure what we value and how we define teaching effectiveness and will give mean-ingful feedback for improvement of instruction

DESIGN PROCESS CHECKLIST

18

Page 19: OEA's Teacher Evaluation & Support System Guidebook

19

Measures are valid (they measure what they intend to)

Measures are reliable (they measure what they intend to measure over a period of time and under similar conditions)

Measures are comparable (they measure the same defini-tion of teaching effectiveness across different grade levels, schools, content areas, etc.)

Weigh measures fairly: ¢Consider the validity/reliability/comparability of a mea-

sure and weigh those more heavily that have the most relevancy and research to back them

¢ Consider level of experience and job description ¢ Consider evaluator capacity (human and resource ca-

pacity strengths and limitations) ¢ Consider that after a pilot, or as a Teacher Evaluation

and Support System progresses, weights of measures can be changed as some are found to be more reliable, valid, comparable, and relevant than others

¢ Consider aligning weights to district priorities, like col-laboration between teachers

¢ Consider that if the ultimate goal is to increase teacher capacity to implement evidence-based practices which will improve student learning, growth and achieve-ment, the observation instrument may need to carry more weight

¢ Consider phasing in reliance on new observation in-struments

Measures support your Teacher Evaluation and Support Sys-tem goals and purposes

Measures have a demonstrated impact on teacher practice

* For ODE Pilots, clearly describe how student growth will be used as a “significant factor.” Include, as a significant factor, at least two student learning goals* (see Student Learning Goals section)¢ Best practice — student performance is based on mul-

tiple valid measures ¢ Determine what are meaningful, valid, reliable and

comparable measures¢ Determine how different student measures apply

across assignments¢ Student data measures should:« Have the ability to accurately measure student

progress between two points in time (growth)« Be rigorous« Be comparable across classrooms

¢ Measures have a demonstrated impact on student achievement

Determine what is required to sustain and support the col-lection, interpretation and application of these multiple valid measures of both teacher and student performance

Initiate a conversation with your bargaining team (consult UniServ Consultant and Center for Great Public Schools) if the district intends to link individual teacher data with their students’ data including unique identifiers for both students and teachers

◗ KEY QUESTIONS TO ASK:

How do you envision multiple valid measures of teaching effectiveness?

How do you ensure that the measures meet the following criteria:¢ Validity¢ Reliability¢ Comparability¢ Relevancy/Meaningfulness

What tradeoffs will the Teacher Evaluation Design Team consider between the comprehensiveness/fidelity of the measures and the practicality (doability/tenability) of the system?¢ What will be the consequences of these tradeoffs?

What does including multiple valid measures of student growth/learning look like in your school district?

When considering tools to assess teachers’ attainment of standards of effective teaching (measures of practice) and tools to assess teachers’ impact on student outcomes (measures of effects), ask yourself these questions:¢ What measures are in place already and how chal-

lenging would a change in evaluation measures be for both evaluators and teachers?

¢ What resources are needed to monitor and sustain the effectiveness of the measure?

Additional questions to consider when developing or se-lecting measures of student learning:¢ Does the approach allow for the assessment of stu-

dent learning over time?¢ How can we establish the following:«The proposed measure assesses the expected

knowledge and skill appropriately, in terms of the content of questions or tasks included and the coverage of the subject area.

« Some students are not disadvantaged by the spe-cific questions or tasks included.

II.Tools

19

Page 20: OEA's Teacher Evaluation & Support System Guidebook

20

¢ The measure appropriately distinguishes among stu-dents.

¢ Scores based on the measure accurately reflect meaning-ful changes in student learning in the subject area, either in strictly comparative terms (e.g., some students learned more than other students) or growth toward a standard (e.g., some student made more progress than others to-ward a goal that will help them be successful).

¢ How can we ensure that student learning is being measured consistently across classrooms/building/district?

¢ What steps in development and administration are needed to ensure that scores will have the same meaning within the subject area and that student growth will have a similar interpretation across sub-jects? (Example, does growth of 15% in mathematics mean the same thing as in English Language arts, and is it achievable across subjects?)

¢ Is the approach and measure transparent and under-standable to stakeholders?

¢ What capacity/resources are needed to develop and implement the measures now and over time?

¢ How easily can data from the model be used along with other data to assess teaching effectiveness?

¢ What provisions are in place to ensure ongoing re-view, calibration, and adjustments, when necessary?

¢ Does the district have a plan to ensure data accuracy?¢ Does the district intend to link individual teacher data

with their students’ data, including unique identifiers for both students and teachers?

¢ What teacher evaluation tools/rubrics will meet both the adopted state standards and the evaluation sys-tem you are developing for your school district?

Does the tool provide multiple valid measures or sources of evidence of teaching practice and effectiveness?

Has the tool been pilot-tested in the field and been shown to have potential to be valid and reliable (aka, it measures what it’s supposed to measure time and time again)?

Does the tool have professional credibility (aka, devel-oped with teacher input, measures important aspects of teaching practice and effectiveness)?

Does the tool provide feedback that teachers can use to improve their practice?

*Does the tool effectively differentiate among four levels of teaching practice and effectiveness that meet the state criteria?*

Who will be evaluated using the proposed system and will different measures be used for different staff?

Given time and resource constraints, are the tools practi-cal? What tradeoffs between practicality and comprehen-siveness must be made?

Will the ratings from each of these measures be weighted to tally a final evaluation rating?

Will classroom observation tools employ checklists, ru-brics, or narratives?

SPECIFIC TO OBSERVATIONS:

◗ Selecting a Rubric:

Will the same rubric be used for all teachers?

◗ Selecting and Training Evaluators:

How will evaluators be selected?

What training will evaluators receive?

Will evaluators be required to demonstrate competency and inter-rater reliability before administrating evaluations?

◗ Conducting Observations:

How many observations will be required?¢ Will this number differentiate between levels of experi-

ence/assignment?

When will the observations take place so that they are ef-fective (e.g., not the last two weeks of school)?

What types of observations (formal and informal) and how many of each will take place and how long will each obser-vation be?

*What will be the pre- and post-observation meeting proto-cols for formal observations?*

◗ Collecting Information:

What information will be collected to support observation findings?

How will the results be shared with the teachers?

How will the district ensure that the results are valid and reliable?

How will disputes of findings be processed?

◗ Refining the Process:

How will the evaluator training be monitored?

How will the teachers be trained on the system and how will this training is monitored?

How will inter-rater reliability be monitored and by whom?

20

DESIGN PROCESS CHECKLIST

Page 21: OEA's Teacher Evaluation & Support System Guidebook

21

How will this information be used to further refine the evalu-ation system?

4. *Professional Development and Growth System (PDGS) that is informed by teacher evaluation* (see page 49 for more) Define how mentoring or peer assistance is used (if at all)

Determine how our Professional Development and Growth System is elevating teaching across the district

Define how the Professional Development and Growth Sys-tem will be aligned to meet the needs of all teachers at all performance levels and provided in a manner that is sup-ported in research

Develop opportunities for teachers to improve that are em-bedded in the evaluation cycle (see CD-ROM for Learning Forward’s Professional Development Standards)

Develop supports to be provided to assist teachers with lower-levels/unacceptable performance

Evaluate outcomes from Teacher Evaluation and Support System to determine whether the Professional Development and Growth System is elevating practice: ¢ Each PD activity¢ Longitudinal analysis of teacher participation, support,

and outcomes related to student learning/growth¢ Ensure that evaluation framework can identify and

provide targeted PD for Special Education and English Language Development personnel

¢ *Develop policy on how Teacher Evaluation and Sup-port System will inform personnel decisions consistent with the collective bargaining agreement and bargain-ing requirements*

◗ KEY QUESTIONS TO ASK:

Does the process provide opportunities for teachers to seek and receive professional development that aligns with their evaluation?

Does the process provide an inspiring roadmap for teacher growth and development?

Does the process support relevant, robust, and timely pro-fessional assistance when needed?

Envision an evaluation system in your district that includes a formative growth process:¢ What would these formative assessments look like?

¢ How would these assessments inform teacher growth and development?

¢ Who would conduct these assessments?¢ How would the “assessors” and “assessees” be

trained on the formative growth process?¢ What criteria should be included?¢ How would these assessments align to/inform the

PGDS?¢ What would this kind of system cost in time and

money?¢ How would it be applied at different venues in the

district?¢ If you could pilot it where should you begin?

Envision how your school district could use the evalua-tion system to align to/inform a professional growth sys-tem in the district:¢ What would that look like?¢ How could this be aligned with curricular standards,

district goals, school goals, and identified needs of students and teachers?

¢ How could this be an application of the InTASC Stan-dards?

¢ What tools/systems would need to be in place in or-der to do this?

¢ What resources would need to be in place?¢ What targeted support would need to be in place?¢ How could you assess the outcomes of the profes-

sional development (PD)?¢ How could the evaluation system and PD systems

support teachers’ new knowledge and skills?¢ If the results of formative assessments are positive,

how would that impact the PD opportunities?¢ If the results of formative assessments identify sig-

nificant shortcomings, how would that impact the PD opportunities?

¢ How can these PD opportunities tie into licensure re-quirements of the state?

¢ How can we individualize PD opportunities and still have a fluency in our overall PGDS?

How will the measures assist in the development of spe-cific performance goals and targeted PD?

Can teacher application and reflection be built into the professional learning activity?

Are professional learning activities “job-embedded” or a one-time-only session? What are the human and fiscal resources that can be used to provide both types of PD?

21

II.Tools

Page 22: OEA's Teacher Evaluation & Support System Guidebook

22

Do teachers have common planning times to reflect upon new practices?

Can opportunities for teachers to observe effective teach-ers be provided?

Will professional learning communities be established?

5. Reworking or revising of current job descriptions and performance standards as necessary

◗ KEY QUESTIONS TO ASK:

Does your new Teacher Evaluation and Support System require reworking or revising current job descriptions and performance standards?

6. Training of both administrators and teachers in the new system Consider the training needs for each of the multiple valid

measures chosen

Establish evaluator training that includes explicit training on measures designed to assess Special Education and English Language Development personnel

Consider that implementation fidelity is most important when the selected measures are dependent on human scoring with observation instruments or rubrics.

Create explicit decision rules and requirements for quan-tity and quality of examples of evidence for making a justi-fication of one performance rating over another which may differentiate for different assignments, experience levels, etc.

*Design a plan for training, due by July 1, 2013 to ODE*

◗ Selecting Evaluators¢ Determine what criteria will be used to select evalu-

ators¢ Determine who will be eligible to conduct the evalu-

ations¢ Consider that not all administrators should necessar-

ily be evaluators – targeted selection may be a better option.

¢ Consider having trained evaluators with knowledge of specialist roles and subject-matter competence

◗ Training Evaluators¢ Create a training system that is co-led by association

leaders and members/teachers designated by the as-sociation, along with administrators

¢ Decide how fidelity will be ensured prior to the pilot, during the pilot, and through-out the life of imple-mentation

¢ Decide how inter-rater reliability and calibration will be ensured prior to the pilot, during the pilot, and through-out the life of implementation

¢ Determine what specialized training for the evalua-tion of or review of specific content or specialty area teachers is needed

¢ Decide how and to what extent the training provided will create opportunities for guided practice

¢ Decide how and to what extent the training will pro-vide specific feedback to improve reliability

¢ Determine what mechanisms will be in place to re-train evaluators/reviewers who are not implementing the system with fidelity

¢ Determine a schedule of regular monitoring of evalu-ators for inter-rater reliability and overall system fi-delity

¢ Determine criteria by which an evaluator will be ter-minated from their evaluation duties

¢ Decide how evaluation responsibilities are a compo-nent of administrative evaluations and that admin-istrators are evaluated on their ability to evaluate teachers with fidelity and reliability

◗ Additional Considerations for Training:¢ Teacher preparation for measuring student learning

growth is limited or non-existent¢ Most principals, support providers, instructional

managers, and coaches are poorly prepared to make judgments about teachers’ contribution to student learning growth

◗ They need to know how to:¢ Evaluate the appropriateness of various measures of

student learning for use in teacher evaluation¢ Work closely with teachers to select appropriate stu-

dent growth measures and ensure that they are using them correctly and consistently.

22

DESIGN PROCESS CHECKLIST

Page 23: OEA's Teacher Evaluation & Support System Guidebook

23

◗ KEY QUESTIONS TO ASK:

How can administrators be trained, supported and moni-tored to carry out an enhanced system?

How is the association involved in the trainings? Co-trainers?

How are all stakeholders trained in the system?

7. Pilot & Implementation or roll-out plan Explore any potential conflicts between the proposed pilot

and the collective bargaining agreement. If necessary, de-velop MOUs that will allow the pilot to move forward with a shared agreement.

Consider ways to either pilot different building levels of the plan or scaffold the district phase-in from the less to more complex component parts

Hold teachers harmless during the pilot years of the Teacher Evaluation and Support System until a measure of validity and reliability can be guaranteed, as well as the degree of relevancy/meaningfulness established

Decide how long the Teacher Evaluation and Support Sys-tem pilot will be and the resources necessary to progress

Decide how data will be taken during the Teacher Evalua-tion and Support System pilot, the frequency and who/how it will be analyzed

Decide the frequency with which modifications will be made and how they will be trained and communicated

Determine whether research will be conducted in conjunc-tion with implementation to provide validation

Establish outcomes/goals to determine the overall effective-ness of the Teacher Evaluation and Support System

Decide if there are resources available to conduct an exter-nal or internal assessment of the evaluation

Rigorously define feedback loop during implementation so details can be assessed and reworked as it rolls out

*Design a plan/timeline for pilot implementation, due to ODE by July 1, 2013*

◗ KEY QUESTIONS TO ASK / QUESTIONS THE PILOT SHOULD ADDRESS:

What criteria will the Teacher Evaluation Design Team create for schools to qualify as a pilot site for the evalua-tion model?

How long will the pilot last?

Are there any conflicts between the proposed evaluation tools and the collective bargaining agreement? If so, what challenges are likely to arise in the negotiation process when labor and management attempt to resolve them?

Who will train the teachers, administrators, and evalua-tors on the new evaluation tools? How will the union make sure that the training addresses reliability issues and equips evaluators to evaluate specific content areas and specialist seffectively? Will evaluators be subject to frequent review or to ensure reliability?

Who will serve as evaluators? What will the criteria be? How will their work be supported and funded? How is their role as an evaluator part of their evaluation system?

When will the training occur? Will compensation be pro-vided to teachers to complete training after the school date or will substitute teacher coverage be provided to complete it during the school day?

How will the evaluation data collected during the pilot be used? Will the stakeholders be “held harmless” during the pilot?

Which measures will be used for formative purposes and which will be used for summative purposes? How will the determination be made?

Will feedback be collected systematically at each stage of the pilot so that changes to the Teacher Evaluation and Support System can be made accordingly? How will that happen? Will changes be made to the Teacher Evaluation and Support System part way through the pilot stage, or will all changes be made after the pilot but before the re-forms are launched district-wide?

◗ QUESTIONS TO ASK STAFF DURING THE PILOT: Feedback from teachers, principal, and other staff members should be collected throughout the pilot to ensure that profes-sional growth is targeted by and improved with the new Teacher Evaluation and Support System. Some questions to ask of staff during the pilot include:

23

II.Tools

Page 24: OEA's Teacher Evaluation & Support System Guidebook

24

Do teachers believe that the new system leads to more tar-geted PD that can improve their performance?

Do individual teachers get to set goals in the new system?

Do teachers feel they have adequate information about the new Teacher Evaluation and Support System? Do they know where to go with questions?

Do teachers believe the new system is fair?

Do teachers understand what the expectations and criteria are for performance at each level, and what it is they need to do in order to progress along the continuum?

Is there anything about the new tools/measures that is confusing to evaluators? To teachers? Do evaluators feel adequately trained to use these new tools? Do teachers?

How will communication be handled between the Associa-tion and members during the development and piloting of the system?

◗ QUESTIONS TO ASK AFTER THE PILOT: Review the feedback from the pilot and ask yourself (and those with whom you are bargaining or jointly developing the system) the following questions:

Which aspects of the Teacher Evaluation and Support Sys-tem are not working as intended? Can they be modified or do they need to be replaced?

Are there aspects of our Teacher Evaluation and Support System that were not foreseen or planned for, which need to be discussed by the Teacher Evaluation Design Team and have a plan developed?

Is there any indication from the pilot that the evaluation results are influencing the district-wide Professional Devel-opment and Growth System planning?

In practice, does the Teacher Evaluation and Support Sys-tem align with InTASC Teaching Standards? Are all tools/measures an accurate application of the Standards?

Are instructional coaches or other resources available to support professional development needs identified in the evaluation results?

Will there be adequate resources/funding to support wide-spread improvements to professional development based on evaluation?

What are the transitions needed to progress to the next stage (additional piloting, initial implementation, etc.)?

8. Evaluation and continuous improvement plan Make sure that a balanced team of teacher and district rep-

resentatives are always a part of evaluating this system throughout its lifetime (contractually protect)

Design an evaluation of your Teacher Evaluation and Sup-port System that can determine whether or not:¢ Stakeholders value and understand the system¢ Student learning/growth is improved¢ Teacher practice is positively impacted¢ Teacher retention is improved¢ The system is implemented with fidelity

Establish a plan to evaluate measures to determine if they can effectively differentiate among teacher performance

Need to identify potential “widget effects” in measures

If measure is not differentiating among teacher practice, may be faulty training or poor implementation, not neces-sarily the measure itself or the teacher/practitioner

Examine correlations among results from measures

Establish a plan/schedule to evaluate processes and data each year and make needed adjustments

◗ KEY QUESTIONS TO ASK:

To what extent does the Teacher Evaluation and Support System (TESS) assess what is under the direct control of teachers?

Does your TESS determine whether or not:¢ Stakeholders value and understand the system?¢ Student learning/growth is improved?¢ Teacher practice is positively impacted?¢ Teacher retention is improved?¢ The system is implemented with fidelity?

What implications does the TESS have for induction?

What implications does the TESS have for recruitment?

What resources are needed to monitor and sustain the ef-fectiveness of the Teacher Evaluation and Support System?¢ How do we avoid having teacher evaluation being just

another education “fad?”¢ How do we continue to prioritize this work?

DESIGN PROCESS CHECKLIST

24

Page 25: OEA's Teacher Evaluation & Support System Guidebook

25

In December 2011, the Oregon State Board of Education adopted national Model Core Teaching Standards,

known as the InTASC Standards. Cre-ated by the Council of Chief State School Officer’s (CCSSO) Interstate Teacher Assessment and Supports Consortium (InTASC), these core teaching standards outline what teachers should know and be able to do. Teacher Evaluation and Sup-port Systems (TESS) in Oregon should be collaboratively customized as an appli-cation of the standards by July 1, 2013. These standards serve as Oregon’s defini-tion of effective teaching and also outline the common principles and foundations of teaching practice that cut across all subject areas and grade levels.

Additionally, these standards are meant to complement and elevate the tremen-dous work occurring in Oregon’s class-rooms and schools on a daily basis, and will continue to foster the growth and professional development of educators. These standards also assume a newer role for teachers as the profession evolves, including responsibilities for facilitating a professional collaborative culture that has both implicit and explicit implications for leadership roles and responsibilities.

InTASC’s TEN STANDARDSInTASC’s ten Standards are divided into four categories: The Learner and Learn-ing, Content Knowledge, Instructional

The three categories of indicators within the InTASC Stan-dards (performances, essential knowledge, and critical dispositions) as a whole are not meant to be copied and

pasted into a checklist or rubric. Essential knowledge and critical dispositions are crucial aspects of teaching effectiveness. They are not aspects that can be readily measured via observations, rubrics, or other measures of teaching. These standards should not be repurposed as a checklist or rubric. Rather, they are the start-ing point for the local customization in which, through the col-laborative process discussed in this guidebook, a comprehensive evaluation system is developed that supports improved teaching effectiveness and student learning.

As you familiarize yourself with these standards in the creation of your TESS, keep in mind that while each standard emphasizes

a discrete aspect of teaching, teaching and learning are infinitely dynamic, integrated and reciprocal processes. In order to fully capture this vibrant relationship, the standards overlap and must be taken as a whole in order to convey a complete picture of the acts of teaching and learning. Also, it is important to keep in mind that indicators are examples of how a teacher might demonstrate each standard.

The indicators used in the standard are simply examples of how a teacher might demonstrate proficiency for each standard; there may be other better indicators in a particular classroom. In other words, an evaluator should not expect a teacher to demonstrate each indicator during a performance assessment.

Check out the 10 standards on the next

page!

Practice, and Professional Responsibility. Each of the ten standards is further sub-divided into three categories of indica-tors: performances, essential knowledge, and critical dispositions. “Performances” are the actions and teaching behaviors that can be observed and assessed, “es-sential knowledge” signals the role of declarative and procedural knowledge as necessary for effective practice, and “criti-cal dispositions” indicates that habits of professional action and moral commit-ments that underlie the performances play a key role in how teachers do, in fact, act in practice.

Getting to Know the InTASC Standards

Standards III.

25

Page 26: OEA's Teacher Evaluation & Support System Guidebook

26

THE LEARNER AND LEARNINGStandard #1: Learner DevelopmentThe teacher understands how students grow and develop, recognizing that pat-terns of learning and development vary individually within and across the cogni-tive, linguistic, social, emotional, and physical areas, and designs and imple-ments developmentally appropriate and challenging learning experiences.

Standard #2: Learning DifferencesThe teacher uses understanding of indi-vidual differences and diverse cultures and communities to ensure inclusive learning environments that enable each student to meet high standards.

Standard #3: Learning EnvironmentsThe teacher works with other to create environments that support individual and collaborative learning, and that encourage positive social interaction, active engage-ment in learning, and self-motivation.

CONTENT KNOWLEDGEStandard #4: Content KnowledgeThe teacher understands the central concepts tools of inquiry, and structures of the discipline(s) he or she teachers and creates learning experiences that make these aspects of the discipline accessible and meaningful for students to assure mastery of the content.

Standard #5: Application of ContentThe teacher understands how to connect concepts and use differing perspectives to engage students in critical thinking, creativity, and collaborative problem solv-ing related to authentic local and global issues.

INSTRUCTIONAL PRACTICEStandard #6: AssessmentThe teacher understands and uses mul-tiple methods of assessment to engage students in their own growth, to monitor learner progress, and to guide the teach-er’s and student’s decision making.

Standard #7: Planning for InstructionThe teacher plans instruction that sup-ports every student in meeting rigorous learning goals by drawing upon knowl-edge of content areas, curriculum, cross disciplinary skills, and pedagogy, as well as knowledge of students and the commu-nity context.

Standard #8: Instructional StrategiesThe teacher understands and uses a variety of instructional strategies to encourage students to develop deep understanding of content areas and their connections, and to build skills to apply knowledge in meaningful ways.

PROFESSIONAL RESPONSIBILITYStandard #9: Professional Learning and Ethical PracticeThe teacher engages in ongoing profes-sional learning and uses evidence to continually evaluate his/her practice, particularly the effects of his/her choices and actions on others (students, families, other professionals, and the community), and adapts practice to meet the needs of each student.

Standard #10: Leadership and CollaborationThe teacher seeks appropriate leadership roles and opportunities to take responsi-bility for student learning, to collaborate with students, families, colleagues, other school professionals, and community members to ensure student growth, and to advance the profession.

THE INTASC TEACHING STANDARDS AT A GLANCE

StandardsIII.

Page 27: OEA's Teacher Evaluation & Support System Guidebook

27

MEA

SURE

S OF

PRO

FESS

IONA

L PRA

CTIC

E

Mea

sure

/ D

escr

iptio

n /

Exam

ples

Rese

arch

Stre

ngth

sCa

utio

nsM

easu

re:

Clas

sroo

m O

bser

vatio

n In

stru

men

ts

Des

crip

tion:

Clas

sroo

m o

bser

vatio

ns a

re th

e m

ost c

omm

on m

easu

re fo

r te

ache

r eva

luat

ion;

they

are

use

d to

mea

sure

obs

erva

ble

clas

sroo

m p

roce

sses

, inc

ludi

ng s

peci

fic te

ache

r pra

ctic

es,

larg

er-p

ictu

re in

stru

ctio

nal a

spec

ts, i

nter

actio

ns b

etw

een

teac

hers

and

stu

dent

s; th

ey v

ary

wid

ely

in h

ow th

ey a

re

cond

ucte

d an

d w

hat t

hey

asse

ss; c

an m

easu

re b

road

er a

s-pe

cts

of te

achi

ng o

r mor

e co

nten

t-/c

onte

xt-s

peci

fic a

spec

ts

as w

ell;

high

-qua

lity

clas

sroo

m o

bser

vatio

n in

stru

men

ts

are

stan

dard

s ba

sed

and

cont

ain

wel

l-spe

cifie

d ru

bric

s th

at

delin

eate

con

sist

ent a

sses

smen

t crit

eria

for e

ach

stan

dard

of

pra

ctic

e; tr

aine

d ev

alua

tors

use

a c

lass

room

obs

erva

-tio

n in

stru

men

t to

mak

e co

nsis

tent

judg

men

ts o

f tea

cher

s’

prac

tice

in th

e cl

assr

oom

; the

obs

erva

tion

inst

rum

ent s

houl

d in

corp

orat

e ev

iden

ce fr

om p

re- a

nd p

ost-

obse

rvat

ion

conf

er-

ence

s w

ith th

e te

ache

r; ob

serv

atio

n in

stru

men

ts a

lso

may

be

used

to a

sses

s te

ache

r pra

ctic

e vi

a vi

deos

of t

heir

less

ons

Exam

ples

:

• Be

aver

ton

Scho

ol D

istr

ict,

Sprin

gfiel

d Sc

hool

Dis

tric

t (se

e C

D-R

OM

for m

ore

Ore

gon

exam

ples

)

• Ch

arlo

tte

Dan

iels

on’s

Fram

ewor

k fo

r Tea

chin

g

• Ci

ncin

nati’

s Te

ache

r Eva

luat

ion

Syst

em

•Tea

cher

Adv

ance

men

t Pro

gram

pro

toco

l

• Cl

assr

oom

Ass

essm

ent S

corin

g Sy

stem

(CLA

SS)

• Pr

otoc

ol fo

r Lan

guag

e A

rts

Teac

hing

Obs

erva

tions

(PLA

TO)

• Ki

m M

arsh

all’s

Rub

rics

• M

arza

no’s

Syst

em

• N

atio

nal C

ompr

ehen

sive

Cen

ter f

or T

each

er Q

ualit

y - “

Prac

-tic

al G

uide

to D

esig

ning

Com

preh

ensi

ve T

each

er E

valu

atio

n Sy

stem

s”

Som

e hi

ghly

rese

arch

ed p

roto

cols

ha

ve b

een

foun

d to

link

to s

tude

nt

achi

evem

ent,

thou

gh a

ssoc

iatio

ns

are

som

etim

es m

odes

t. Re

sear

ch a

nd

valid

ity fi

ndin

gs a

re h

ighl

y de

pend

ent

on th

e in

stru

men

t use

d, s

ampl

ing

proc

edur

es, a

nd tr

aini

ng o

f rat

ers.

Th

ere

is a

lack

of r

esea

rch

on o

bser

-va

tion

prot

ocol

s as

use

d in

con

text

fo

r tea

cher

eva

luat

ion.

To b

e su

cces

sful

:

• N

eeds

to b

e fre

quen

t (ev

ery

2-3

wee

ks)

• Co

mbi

natio

n of

form

al a

nd in

form

al

• C

andi

d, e

vide

nce

base

d fe

edba

ck is

ne

eded

• Pr

ofes

sion

ally

cre

dibl

e as

they

ofte

n ar

e de

vel-

oped

by

teac

hers

and

exp

erts

in te

ache

r edu

catio

n

• C

an b

e ad

apte

d fo

r var

ious

sub

ject

s, g

rade

s, a

nd

cont

exts

• C

an m

easu

re m

any

aspe

cts

of p

ract

ice

that

are

as

soci

ated

with

effe

ctiv

e te

achi

ng

• U

sual

ly b

ased

on

vett

ed s

tand

ards

of p

rofe

s-si

onal

pra

ctic

e

• So

me

obse

rvat

ion

inst

rum

ents

hav

e be

en d

em-

onst

rate

d to

be

rela

ted

to s

tude

nt a

chie

vem

ent

• C

an p

rovi

de u

sefu

l for

mat

ive

and

sum

mat

ive

info

rmat

ion

• Pr

ovid

es ri

ch in

form

atio

n ab

out c

lass

room

be-

havi

ors

and

activ

ities

• Is

gen

eral

ly c

onsi

dere

d a

fair

and

dire

ct m

easu

re

by s

take

hold

ers

• C

aref

ul a

tten

tion

mus

t be

paid

to

choo

sing

or c

reat

ing

a va

lid a

nd re

liabl

e pr

otoc

ol a

nd tr

aini

ng a

nd c

alib

ratin

g ra

ters

•Hig

h qu

ality

, effe

ctiv

e cl

assr

oom

obs

er-

vatio

n th

at is

val

id a

nd re

liabl

e re

quire

s ex

tens

ive

trai

ning

of t

he e

valu

ator

s an

d ca

n be

exp

ensi

ve a

nd ti

me-

cons

umin

g;

inte

nsiv

e tr

aini

ng a

nd c

alib

ratin

g of

ob-

serv

ers

adds

to e

xpen

se b

ut is

nec

essa

ry

for v

alid

ity

• Th

is m

etho

d as

sess

es o

bser

vabl

e cl

assr

oom

beh

avio

rs b

ut is

not

as

usef

ul

for a

sses

sing

bel

iefs

, fee

lings

, int

entio

ns,

or o

ut-o

f-cl

assr

oom

act

iviti

es

• M

any

obse

rvat

ion

inst

rum

ents

not

va

lidat

ed in

term

s of

bei

ng d

emon

stra

ted

to b

e re

late

d to

stu

dent

lear

ning

• A

sses

ses

teac

her p

ract

ice

but n

ot e

f-fe

cts

on s

tude

nt le

arni

ng

• Te

ache

rs a

re o

ften

left

out o

f any

tr

aini

ng re

late

d to

the

eval

uatio

n in

stru

-m

ent,

whi

ch in

clud

es a

sha

red

visi

on o

f eff

ectiv

e te

achi

ng a

nd s

hare

d la

ngua

ge

desc

ribin

g th

e vi

sion

• So

me

live

vide

o-st

ream

ing

of c

lass

-ro

oms

do n

ot in

corp

orat

e pr

e- a

nd p

ost-

obse

rvat

ion

conf

eren

ces.

In th

ese

situ

a-tio

ns, o

bser

vers

do

not h

ave

the

bene

fit

of a

sses

sing

the

teac

hing

in c

onte

xt

• In

flate

d ra

tings

• C

an b

e in

frequ

ent

• C

ritic

ism

can

be

subj

ectiv

e

IV.

This

cha

rt p

rovi

des

you

with

sev

eral

exa

mpl

es o

f mea

sure

s of

teac

hing

effe

ctiv

enes

s w

ithin

two

of th

e th

ree

“buc

kets

” of

mea

sure

s of

evi

denc

e re

quire

d by

the

ESEA

Wai

ver:

Mea

sure

s of

Pro

fess

iona

l Pra

ctic

e an

d M

easu

res

of P

rofe

ssio

nal R

espo

n-si

bilit

y. F

or e

ach

mea

sure

, the

re is

a d

escr

iptio

n w

ith e

xam

ples

, cu

rren

t res

earc

h, a

nd a

list

of t

he m

easu

re’s

stre

ngth

s as

wel

l as

cau

tions

for u

se.

Wha

t You

're Lo

okin

g At

...

Page 28: OEA's Teacher Evaluation & Support System Guidebook

28

Mea

sure

:

Non

-cla

ssro

om o

bser

vatio

ns

Des

crip

tion:

Obs

erva

tions

of n

on-c

lass

room

set

tings

, with

out s

tude

nts,

bu

t can

incl

ude:

sta

ff m

eetin

gs, p

aren

t mee

tings

, PLC

s/co

l-la

bora

tive

mee

tings

, cur

ricul

um d

esig

n m

eetin

gs, p

repa

ratio

n pe

riods

, et

c

Not

man

y ex

plic

it re

sear

ch s

tudi

es o

n no

n-cl

assr

oom

obs

erva

tion

only

.•

Prof

essi

onal

ly c

redi

ble

as th

ey o

ften

are

deve

l-op

ed b

y te

ache

rs a

nd e

xper

ts in

teac

her e

duca

tion

• C

an b

e ad

apte

d fo

r var

ious

con

text

s

• C

an m

easu

re m

any

aspe

cts

of p

ract

ice

that

are

as

soci

ated

with

effe

ctiv

e te

achi

ng

• C

an b

e ba

sed

on v

ette

d st

anda

rds

of p

rofe

ssio

nal

prac

tice

(InT

ASC

)

• C

an p

rovi

de u

sefu

l for

mat

ive

and

sum

mat

ive

info

rmat

ion

• Is

gen

eral

ly c

onsi

dere

d a

fair

and

dire

ct m

easu

re

by s

take

hold

ers

• C

aref

ul a

tten

tion

mus

t be

paid

to

choo

sing

or c

reat

ing

a va

lid a

nd re

liabl

e pr

otoc

ol a

nd tr

aini

ng a

nd c

alib

ratin

g ra

ters

•Hig

h qu

ality

, effe

ctiv

e ob

serv

atio

n th

at

is v

alid

and

relia

ble

requ

ires

exte

nsiv

e tr

aini

ng o

f the

eva

luat

ors

and

can

be

expe

nsiv

e an

d tim

e-co

nsum

ing

• A

sses

ses

teac

her p

ract

ice

but n

ot e

f-fe

cts

on s

tude

nt le

arni

ng

• In

flate

d ra

tings

• C

an b

e in

frequ

ent

• C

ritic

ism

can

be

subj

ectiv

e

Mea

sure

:

Prin

cipa

l Eva

luat

ion

Des

crip

tion:

Is g

ener

ally

bas

ed o

n cl

assr

oom

obs

erva

tion,

may

be

stru

c-tu

red

or u

nstr

uctu

red;

use

s an

d pr

oced

ures

var

y w

idel

y by

di

stric

t. Is

gen

eral

ly u

sed

for s

umm

ativ

e pu

rpos

es, m

ost

com

mon

ly fo

r ten

ure

or d

ism

issa

l dec

isio

ns fo

r beg

inni

ng

teac

hers

.

Stud

ies

com

parin

g su

bjec

tive

prin

-ci

pal r

atin

gs to

stu

dent

ach

ieve

men

t fin

d m

ixed

resu

lts. L

ittle

evi

denc

e ex

-is

ts o

n va

lidity

of e

valu

atio

ns a

s th

ey

occu

r in

scho

ols,

but

evi

denc

e ex

ists

th

at tr

aini

ng fo

r prin

cipa

ls is

lim

ited

and

rare

, whi

ch w

ould

impa

ir va

lidity

of

thei

r eva

luat

ions

.

• C

an re

pres

ent a

use

ful p

ersp

ectiv

e ba

sed

on

prin

cipa

ls’ k

now

ledg

e of

sch

ool a

nd c

onte

xt

• Is

gen

eral

ly fe

asib

le a

nd c

an b

e on

e us

eful

co

mpo

nent

in a

sys

tem

use

d to

mak

e su

mm

ativ

e ju

dgm

ents

and

pro

vide

form

ativ

e fe

edba

ck.

• Ev

alua

tion

inst

rum

ents

use

d w

ithou

t pr

oper

trai

ning

or r

egar

d fo

r the

ir in

-te

nded

pur

pose

will

impa

ir va

lidity

.

• Pr

inci

pals

may

not

be

qual

ified

to e

valu

-at

e te

ache

rs o

n m

easu

res

high

ly s

peci

al-

ized

for c

erta

in s

ubje

cts

or c

onte

xts.

Mea

sure

:

Art

ifact

Ana

lysi

s

Des

crip

tions

:

Thes

e in

stru

men

ts a

nd s

truc

ture

d pr

otoc

ols

used

to a

naly

ze

clas

sroo

m a

rtifa

cts

in o

rder

to d

eter

min

e th

e qu

ality

of

inst

ruct

ion

in a

cla

ssro

om; m

ay ra

te le

sson

pla

ns, t

each

er a

s-si

gnm

ents

, tea

cher

-cre

ated

ass

essm

ents

, and

sco

ring

rubr

ics

on p

artic

ular

crit

eria

, suc

h as

rigo

r, au

then

ticity

, int

elle

ctua

l de

man

d, a

nd a

lignm

ent t

o st

anda

rds,

cla

rity,

and

com

preh

en-

sive

ness

.

Exam

ples

:

• In

stru

ctio

nal Q

ualit

y A

sses

smen

ts (I

QA

)

• Te

ache

r Wor

k Sa

mpl

e M

etho

dolo

gy

Pilo

t res

earc

h ha

s lin

ked

artif

act

ratin

gs to

obs

erve

d m

easu

res

of

prac

tice,

qua

lity

of s

tude

nt w

ork,

and

st

uden

t ach

ieve

men

t gai

ns. M

ore

wor

k is

nee

ded

to e

stab

lish

scor

ing

relia

bilit

y an

d de

term

ine

the

idea

l am

ount

of w

ork

to s

ampl

e. L

ack

of

rese

arch

exi

sts

on u

se o

f str

uctu

red

artif

act a

naly

sis

in p

ract

ice.

• Pr

ofes

sion

ally

cre

dibl

e •

Ada

ptab

le fo

r diff

eren

t typ

es o

f tea

cher

s •

Non

inva

sive

, doe

s no

t nee

d to

be

done

in re

al

time

• C

aptu

res

man

y as

pect

s of

teac

her p

ract

ice

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

form

atio

n on

teac

her p

ract

ice

• C

an b

e a

usef

ul m

easu

re o

f ins

truc

tiona

l qua

lity

if a

valid

ated

pro

toco

l is

used

, if r

ater

s ar

e w

ell-

trai

ned

for r

elia

bilit

y, a

nd if

ass

ignm

ents

sho

w

suffi

cien

t var

iatio

n in

qua

lity

• Is

pra

ctic

al a

nd fe

asib

le b

ecau

se a

rtifa

cts

have

al

read

y be

en c

reat

ed fo

r the

cla

ssro

om

• Fe

w v

alid

ated

sys

tem

s ex

ist

• Co

mpa

rabi

lity

acro

ss d

iffer

ent t

ypes

of

teac

hers

has

not

bee

n es

tabl

ishe

d •

Relia

bilit

y ac

ross

obs

erve

rs re

quire

s ex

tens

ive

trai

ning

Doe

s no

t inc

lude

mea

sure

s of

stu

dent

le

arni

ng

• M

ore

valid

ity a

nd re

liabi

lity

rese

arch

is

need

ed

• Tr

aini

ng k

now

ledg

eabl

e sc

orer

s ca

n be

co

stly

but

is n

eces

sary

to e

nsur

e va

lidity

• Th

is m

etho

d m

ay b

e a

prom

isin

g m

iddl

e gr

ound

in te

rms

of fe

asib

ility

and

va

lidity

bet

wee

n fu

ll ob

serv

atio

n an

d le

ss

dire

ct m

easu

res

such

as

self-

repo

rt

Page 29: OEA's Teacher Evaluation & Support System Guidebook

29

MEA

SURE

S OF

PRO

FESS

IONA

L RES

PONS

IBIL

ITIE

S

Mea

sure

/ D

escr

iptio

n /

Exam

ples

Rese

arch

Stre

ngth

sCa

utio

nsM

easu

re:

Teac

her S

elf-

Repo

rt/

Refle

ctio

ns

Des

crip

tion:

Teac

her s

elf-

repo

rt m

easu

res

may

take

the

form

of s

urve

ys, i

nstr

uc-

tiona

l log

s, o

r int

ervi

ews.

Can

var

y w

idel

y in

focu

s an

d le

vel o

f det

ail.

They

as

k te

ache

rs to

repo

rt o

n w

hat t

hey

are

doin

g in

the

clas

sroo

m, t

he

exte

nt to

whi

ch th

ey a

re

mee

ting

stan

dard

s, a

nd in

som

e ca

ses

anal

yze

the

impa

ct o

f the

ir pr

actic

e. T

hey

may

con

sist

of c

heck

lists

, rat

ing

scal

es, r

ubric

s, a

nd

may

requ

ire te

ache

rs to

indi

cate

the

frequ

ency

of p

artic

ular

pra

c-tic

es.

Exam

ples

:

• St

udy

of In

stru

ctio

nal I

mpr

ovem

ent i

nstr

uctio

nal l

ogs

• Se

lf-as

sess

men

ts

• N

atio

nal B

oard

for P

rofe

ssio

nal T

each

ing

Stan

dard

s

Stud

ies

on th

e va

lidity

of t

each

er s

elf-

repo

rt m

easu

res

pres

ent m

ixed

resu

lts.

Hig

hly

deta

iled

mea

sure

s of

pra

ctic

e m

ay

be b

ette

r abl

e to

cap

ture

act

ual t

each

ing

prac

tices

but

may

be

hard

er to

est

ablis

h re

liabi

lity

or m

ay re

sult

in v

ery

narr

owly

fo

cuse

d m

easu

res.

• C

an m

easu

re u

nobs

erva

ble

fact

ors

that

may

affe

ct

teac

hing

, suc

h as

kno

wle

dge,

inte

ntio

ns, e

xpec

tatio

ns,

and

belie

fs

• Pr

ovid

es th

e un

ique

per

spec

tive

of th

e te

ache

r

•Is

very

feas

ible

and

cos

t-effi

cien

t; ca

n co

llect

larg

e am

ount

s of

info

rmat

ion

at o

nce

• C

an m

easu

re u

nobs

erva

ble

aspe

cts

of te

ache

r qua

lity

• C

an b

e ea

sily

adm

inis

tere

d

• C

an p

rom

ote

teac

her s

elf-

refle

ctio

n an

d an

alys

is

• C

an p

rom

ote

a se

nse

of s

elf-

effica

cy

• D

o no

t inc

lude

inde

pend

ent m

easu

res

of im

pact

on

stud

ent l

earn

ing

• C

an p

rovi

de u

sefu

l for

mat

ive

info

rmat

ion

• Re

liabi

lity

and

valid

ity o

f sel

f-re

port

for s

um-

mat

ive

purp

oses

is n

ot fu

lly e

stab

lishe

d an

d de

pend

s on

inst

rum

ent u

sed

• U

sing

or c

reat

ing

a w

ell-d

evel

oped

and

val

i-da

ted

inst

rum

ent w

ill d

ecre

ase

cost

-effi

cien

cy

but w

ill in

crea

se a

ccur

acy

of fi

ndin

gs

• Th

is m

etho

d sh

ould

not

be

used

as

a so

le o

r pr

imar

y m

easu

re in

teac

her e

valu

atio

n

• In

stru

ctio

nal l

ogs

have

bee

n us

ed o

nly

for

rese

arch

pur

pose

s, s

o th

eir v

alid

ity fo

r eva

lua-

tion

is q

uest

iona

ble

• D

o no

t inc

lude

inde

pend

ent m

easu

res

of

impa

ct o

n st

uden

t lea

rnin

g

Mea

sure

:

Goa

l Set

ting

Des

crip

tion:

Thes

e ev

alua

tion

tool

s off

er te

ache

rs th

e op

port

unity

to s

et th

eir o

wn

high

but

feas

ible

obj

ectiv

es fo

r the

ir st

uden

ts’ g

row

th in

col

labo

ra-

tion

with

thei

r prin

cipa

l and

/or o

ther

col

leag

ues.

The

ass

essm

ents

te

ache

rs u

se m

ay b

e co

mm

on o

r sta

ndar

dize

d ex

ams

or te

ache

r-de

velo

ped

asse

ssm

ents

. Som

e to

ols

requ

ire te

ache

rs to

spe

cify

the

prof

essi

onal

dev

elop

men

t the

y w

ill p

artic

ipat

e in

to e

nsur

e th

eir

stud

ents

ach

ieve

thei

r gro

wth

obj

ectiv

es.

Exam

ples

:

• Sp

ringfi

eld

Scho

ol D

istr

ict (

see

CD

-RO

M)

• O

rego

n Fr

amew

ork

(See

CD

-RO

M)

Relia

bilit

y an

d va

lidity

of g

oal-s

ettin

g fo

r su

mm

ativ

e pu

rpos

es is

not

fully

est

ab-

lishe

d

•Is

com

preh

ensi

ve a

nd c

an m

easu

re a

spec

ts o

f tea

chin

g th

at a

re n

ot re

adily

obs

erva

ble

in th

e cl

assr

oom

• C

an b

e us

ed w

ith te

ache

rs o

f all

field

s

•Pro

vide

s a

high

leve

l of c

redi

bilit

y am

ong

stak

ehol

ders

•Is

a go

od to

ol fo

r tea

cher

refle

ctio

n an

d im

prov

emen

t

• Pr

ofes

sion

ally

cre

dibl

e

• C

an b

e do

ne e

lect

roni

cally

• M

ay in

clud

e so

me

asse

ssm

ent o

f stu

dent

lear

ning

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

for-

mat

ion

on te

ache

r pra

ctic

e

•Is

very

feas

ible

and

cos

t-effi

cien

t; ca

n co

llect

larg

e am

ount

s of

info

rmat

ion

at o

nce

• C

an b

e ea

sily

adm

inis

tere

d

• C

an p

rom

ote

a se

nse

of s

elf-

effica

cy

•Diffi

cult

to s

tand

ardi

ze (c

ompa

re a

cros

s te

ache

rs o

r sch

ools

)

• M

ay n

ot in

clud

e ro

bust

mea

sure

s of

stu

dent

le

arni

ng

• Re

liabi

lity

and

valid

ity o

f goa

l-set

ting

for

sum

mat

ive

purp

oses

is n

ot fu

lly e

stab

lishe

d

• Th

is m

etho

d sh

ould

not

be

used

as

a so

le o

r pr

imar

y m

easu

re in

teac

her e

valu

atio

n

Page 30: OEA's Teacher Evaluation & Support System Guidebook

30

Mea

sure

:

Port

folio

s

Des

crip

tion:

Teac

her p

ortfo

lios

are

exhi

bits

of e

vide

nce

of te

achi

ng p

rac-

tice,

sch

ool a

ctiv

ities

, and

stu

dent

pro

gres

s. T

hey

usua

lly

are

com

pile

d by

the

teac

her h

im- o

r her

self.

Por

tfolio

s m

ay

incl

ude

teac

her-

crea

ted

less

on o

r uni

t pla

ns, d

escr

iptio

ns o

f th

e cl

assr

oom

con

text

, ass

ignm

ents

, stu

dent

wor

k sa

mpl

es,

vide

os o

f cla

ssro

om in

stru

ctio

n, n

otes

from

par

ents

, and

te

ache

rs’ a

naly

ses

of th

eir s

tude

nts

lear

ning

in re

latio

n to

th

eir i

nstr

uctio

n. T

hey

can

be u

sed

to d

ocum

ent a

larg

e ra

nge

of te

achi

ng b

ehav

iors

and

resp

onsi

bilit

ies.

Has

bee

n us

ed w

idel

y in

teac

her e

duca

tion

prog

ram

s an

d in

st

ates

for a

sses

sing

the

perfo

rman

ce o

f tea

cher

can

dida

tes

and

begi

nnin

g te

ache

rs.

Rese

arch

on

valid

ity a

nd re

liabi

lity

is o

ngoi

ng, a

nd c

once

rns

have

bee

n ra

ised

abo

ut c

onsi

sten

cy/

stab

ility

in

sco

ring.

The

re is

a la

ck o

f res

earc

h lin

king

por

tfolio

s to

stu

dent

ach

ieve

-m

ent.

Som

e st

udie

s ha

ve li

nked

N

BPTS

cer

tifica

tion

(whi

ch in

clud

es

a po

rtfo

lio) t

o st

uden

t ach

ieve

men

t, bu

t oth

er s

tudi

es h

ave

foun

d no

re

latio

nshi

p.

Exam

ples

:

• N

atio

nal B

oard

for P

rofe

ssio

nal

Teac

hing

Sta

ndar

ds p

ortfo

lio a

sses

s-m

ent

•Tea

cher

Per

form

ance

Ass

essm

ent (

TPA

)

•Per

form

ance

Ass

essm

ent o

f Cal

ifor-

nia

Teac

hers

(PA

CT)

•Is

com

preh

ensi

ve a

nd c

an m

easu

re a

spec

ts o

f te

achi

ng th

at a

re n

ot re

adily

obs

erva

ble

in th

e cl

assr

oom

• C

an b

e us

ed w

ith te

ache

rs o

f all

field

s

•Pro

vide

s a

high

leve

l of c

redi

bilit

y am

ong

stak

e-ho

lder

s

•Is

a go

od to

ol fo

r tea

cher

refle

ctio

n an

d im

prov

e-m

ent

• Pr

ofes

sion

ally

cre

dibl

e

• C

an b

e do

ne e

lect

roni

cally

• M

ay in

clud

e so

me

asse

ssm

ent o

f stu

dent

lear

n-in

g

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

form

atio

n on

teac

her p

ract

ice

•Por

tfolio

s re

pres

ent t

each

ers’

exe

mpl

ary

wor

k

•Sco

rers

sho

uld

have

con

tent

kno

wle

dge

of th

e po

rtfo

lios

•The

sta

bilit

y of

sco

res

may

not

be

high

en

ough

to u

se fo

r hig

h-st

akes

ass

ess-

men

t

•Por

tfolio

s ar

e di

fficu

lt to

sta

ndar

dize

(c

ompa

re a

cros

s te

ache

rs o

r sch

ools

)

• Ti

me-

cons

umin

g fo

r tea

cher

s an

d to

a

less

er e

xten

t sco

rers

• M

ay n

ot re

pres

ent d

ay-t

o-da

y pr

actic

e w

ell

• Po

rtfo

lio s

core

s se

ldom

hav

e be

en

show

n to

be

cons

iste

ntly

rela

ted

to s

tu-

dent

ach

ieve

men

t

• M

ay n

ot in

clud

e ro

bust

mea

sure

s of

st

uden

t lea

rnin

g

Mea

sure

:

Peer

Col

labo

ratio

n &

Ass

ista

nce

(FO

RMAT

IVE

ON

LY a

s on

ly li

cens

ed a

dmin

istr

ator

s ca

n ev

alua

te te

ache

rs p

er th

e Te

ache

r Sta

ndar

ds a

nd P

ract

ices

Co

mm

issi

on (T

SPC)

. TS

PC h

as s

tatu

tory

aut

horit

y to

lic

ense

and

defi

ne th

e ro

les

of A

dmin

istr

ator

s (O

RS 3

42.12

1 an

d O

RS 3

42.14

0) a

nd th

ey h

ave

ther

efor

e de

fined

lice

nsed

A

dmin

istr

ator

s as

the

only

 “te

ache

rs”

allo

wed

to e

valu

-at

e an

d su

perv

ise

(OA

R 58

4-00

5-00

05) w

hich

is fu

rthe

r de

fined

and

rest

ricte

d by

sta

te b

arga

inin

g la

w (P

ublic

Em

ploy

ee C

olle

ctiv

e Ba

rgai

ning

Act

ORS

243

.650

; Pee

r A

ssis

tanc

e al

so m

ust b

e ca

rrie

d ou

t per

OA

R 34

2.85

0 (2

)(b

)(E)

)

Des

crip

tion:

Peer

ass

ista

nce

(PA

) pro

gram

s ar

e jo

int,

colla

bora

tive

labo

r-m

anag

emen

t pro

gram

s th

at fo

cus

on e

nhan

cing

teac

her

qual

ity b

y us

ing

expe

rt te

ache

rs a

s m

ento

rs fo

r pro

batio

nary

or

car

eer t

each

ers.

(See

Eva

luat

ion

Proc

edur

es S

ectio

n fo

r m

ore)

.

Exam

ples

: •

Mas

sach

uset

ts (R

einv

entin

g Ed

ucat

or E

valu

atio

n)**

New

Yor

k (T

each

er E

valu

atio

n an

d D

evel

opm

ent H

andb

ook)

**

**Pl

ease

not

e th

at b

oth

thes

e sy

stem

s in

clud

e Pe

er A

ssis

-ta

nce

and

Revi

ew c

ompo

nent

s an

d th

is c

hart

is o

nly

addr

ess-

ing

the

Peer

Ass

ista

nce

part

Evid

ence

(Hum

phre

y, e

t. al

, p. 2

2)

sugg

ests

that

CTs

tend

to c

ondu

ct

mor

e co

mpr

ehen

sive

ass

essm

ents

th

an p

rinci

pals

who

are

mor

e ty

pica

lly u

sing

a c

heck

list t

o id

entif

y is

sues

that

nee

d at

tent

ion.

CTs

typi

-ca

lly s

pend

mor

e tim

e w

ith te

ache

rs

durin

g PA

or P

AR

than

prin

cipa

ls o

r ad

min

istr

ator

s—an

d th

eir f

eedb

ack

is

ofte

n m

ore

supp

ortiv

e, le

ss th

reat

en-

ing,

and

bro

ader

in s

cope

.

• W

orks

bes

t whe

n de

sign

ed c

olla

bora

tivel

y w

ith

labo

r-m

anag

emen

t sta

keho

lder

s

•Tho

ugh

PA m

ay c

ost m

ore

initi

ally

, man

y sc

hool

di

stric

ts re

port

that

ove

r tim

e, P

A p

rogr

ams

redu

ce

turn

over

cos

ts, s

uppo

rt te

ache

r rec

ruitm

ent a

nd

rete

ntio

n, a

nd d

ecre

ase

the

time

and

expe

nse

as-

soci

ated

with

dis

mis

sing

car

eer t

each

ers

• C

an e

xpan

d ca

paci

ty fo

r for

mat

ive

feed

back

• C

an w

ork

flexi

bly

with

in a

bui

ldin

g or

dis

tric

t’s

uniq

ue s

ched

ules

and

nee

ds

• M

ay fe

el le

ss th

reat

enin

g th

an re

ceiv

ing

feed

back

fro

m p

rinci

pal

• Co

uld

have

mor

e fle

xibi

lity

to h

ave

CT w

ith

sim

ilar e

xper

ienc

es, t

hus

feed

back

wou

ld b

e m

ore

valid

and

relia

ble

• PA

are

exp

ensi

ve in

bot

h tim

e an

d m

oney

to im

plem

ent.

A s

choo

l dis

tric

t m

ust h

ire te

ache

rs to

repl

ace

the

CTs

in

the

clas

sroo

m. O

ther

cos

ts a

re th

e st

i-pe

nds

paid

to C

Ts a

nd g

ener

al a

dmin

is-

trat

ive

and

trai

ning

cos

ts a

ssoc

iate

d w

ith

the

prog

ram

• CT

job

resp

onsi

bilit

ies

are

com

preh

en-

sive

and

dem

andi

ng –

mus

t be

thor

ough

-ly

thou

ght t

hrou

gh a

nd te

nabl

e

• CT

wou

ld n

eed

high

leve

l of t

rain

ing

and

stan

dard

izat

ion

to m

ake

resu

lts c

ompa

-ra

ble,

relia

ble,

and

val

id

• CT

s w

ould

nee

d to

be

car

eful

ly c

hose

n th

roug

h a

high

ly th

ough

tful p

roce

ss

• Th

is s

yste

m m

ust h

ave

stak

ehol

der

buy-

in u

p fro

nt –

this

mea

ns ti

me

and

mon

ey to

trai

n an

d ed

ucat

e al

l sta

keho

ld-

ers

prio

r to

an a

sk o

f buy

-in

• Th

is s

yste

m c

ould

lead

to a

det

erio

ra-

tion

of re

latio

nshi

ps b

etw

een

teac

hers

if

it is

not

impl

emen

ted

with

abu

ndan

t tr

aini

ng, r

esou

rces

, tim

e an

d th

ough

tful-

ness

Page 31: OEA's Teacher Evaluation & Support System Guidebook

31

Mea

sure

:

Port

folio

s

Des

crip

tion:

Teac

her p

ortfo

lios

are

exhi

bits

of e

vide

nce

of te

achi

ng p

rac-

tice,

sch

ool a

ctiv

ities

, and

stu

dent

pro

gres

s. T

hey

usua

lly

are

com

pile

d by

the

teac

her h

im- o

r her

self.

Por

tfolio

s m

ay

incl

ude

teac

her-

crea

ted

less

on o

r uni

t pla

ns, d

escr

iptio

ns o

f th

e cl

assr

oom

con

text

, ass

ignm

ents

, stu

dent

wor

k sa

mpl

es,

vide

os o

f cla

ssro

om in

stru

ctio

n, n

otes

from

par

ents

, and

te

ache

rs’ a

naly

ses

of th

eir s

tude

nts

lear

ning

in re

latio

n to

th

eir i

nstr

uctio

n. T

hey

can

be u

sed

to d

ocum

ent a

larg

e ra

nge

of te

achi

ng b

ehav

iors

and

resp

onsi

bilit

ies.

Has

bee

n us

ed w

idel

y in

teac

her e

duca

tion

prog

ram

s an

d in

st

ates

for a

sses

sing

the

perfo

rman

ce o

f tea

cher

can

dida

tes

and

begi

nnin

g te

ache

rs.

Rese

arch

on

valid

ity a

nd re

liabi

lity

is o

ngoi

ng, a

nd c

once

rns

have

bee

n ra

ised

abo

ut c

onsi

sten

cy/

stab

ility

in

sco

ring.

The

re is

a la

ck o

f res

earc

h lin

king

por

tfolio

s to

stu

dent

ach

ieve

-m

ent.

Som

e st

udie

s ha

ve li

nked

N

BPTS

cer

tifica

tion

(whi

ch in

clud

es

a po

rtfo

lio) t

o st

uden

t ach

ieve

men

t, bu

t oth

er s

tudi

es h

ave

foun

d no

re

latio

nshi

p.

Exam

ples

:

• N

atio

nal B

oard

for P

rofe

ssio

nal

Teac

hing

Sta

ndar

ds p

ortfo

lio a

sses

s-m

ent

•Tea

cher

Per

form

ance

Ass

essm

ent (

TPA

)

•Per

form

ance

Ass

essm

ent o

f Cal

ifor-

nia

Teac

hers

(PA

CT)

•Is

com

preh

ensi

ve a

nd c

an m

easu

re a

spec

ts o

f te

achi

ng th

at a

re n

ot re

adily

obs

erva

ble

in th

e cl

assr

oom

• C

an b

e us

ed w

ith te

ache

rs o

f all

field

s

•Pro

vide

s a

high

leve

l of c

redi

bilit

y am

ong

stak

e-ho

lder

s

•Is

a go

od to

ol fo

r tea

cher

refle

ctio

n an

d im

prov

e-m

ent

• Pr

ofes

sion

ally

cre

dibl

e

• C

an b

e do

ne e

lect

roni

cally

• M

ay in

clud

e so

me

asse

ssm

ent o

f stu

dent

lear

n-in

g

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

form

atio

n on

teac

her p

ract

ice

•Por

tfolio

s re

pres

ent t

each

ers’

exe

mpl

ary

wor

k

•Sco

rers

sho

uld

have

con

tent

kno

wle

dge

of th

e po

rtfo

lios

•The

sta

bilit

y of

sco

res

may

not

be

high

en

ough

to u

se fo

r hig

h-st

akes

ass

ess-

men

t

•Por

tfolio

s ar

e di

fficu

lt to

sta

ndar

dize

(c

ompa

re a

cros

s te

ache

rs o

r sch

ools

)

• Ti

me-

cons

umin

g fo

r tea

cher

s an

d to

a

less

er e

xten

t sco

rers

• M

ay n

ot re

pres

ent d

ay-t

o-da

y pr

actic

e w

ell

• Po

rtfo

lio s

core

s se

ldom

hav

e be

en

show

n to

be

cons

iste

ntly

rela

ted

to s

tu-

dent

ach

ieve

men

t

• M

ay n

ot in

clud

e ro

bust

mea

sure

s of

st

uden

t lea

rnin

g

Mea

sure

:

Peer

Col

labo

ratio

n &

Ass

ista

nce

(FO

RMAT

IVE

ON

LY a

s on

ly li

cens

ed a

dmin

istr

ator

s ca

n ev

alua

te te

ache

rs p

er th

e Te

ache

r Sta

ndar

ds a

nd P

ract

ices

Co

mm

issi

on (T

SPC)

. TS

PC h

as s

tatu

tory

aut

horit

y to

lic

ense

and

defi

ne th

e ro

les

of A

dmin

istr

ator

s (O

RS 3

42.12

1 an

d O

RS 3

42.14

0) a

nd th

ey h

ave

ther

efor

e de

fined

lice

nsed

A

dmin

istr

ator

s as

the

only

 “te

ache

rs”

allo

wed

to e

valu

-at

e an

d su

perv

ise

(OA

R 58

4-00

5-00

05) w

hich

is fu

rthe

r de

fined

and

rest

ricte

d by

sta

te b

arga

inin

g la

w (P

ublic

Em

ploy

ee C

olle

ctiv

e Ba

rgai

ning

Act

ORS

243

.650

; Pee

r A

ssis

tanc

e al

so m

ust b

e ca

rrie

d ou

t per

OA

R 34

2.85

0 (2

)(b

)(E)

)

Des

crip

tion:

Peer

ass

ista

nce

(PA

) pro

gram

s ar

e jo

int,

colla

bora

tive

labo

r-m

anag

emen

t pro

gram

s th

at fo

cus

on e

nhan

cing

teac

her

qual

ity b

y us

ing

expe

rt te

ache

rs a

s m

ento

rs fo

r pro

batio

nary

or

car

eer t

each

ers.

(See

Eva

luat

ion

Proc

edur

es S

ectio

n fo

r m

ore)

.

Exam

ples

: •

Mas

sach

uset

ts (R

einv

entin

g Ed

ucat

or E

valu

atio

n)**

New

Yor

k (T

each

er E

valu

atio

n an

d D

evel

opm

ent H

andb

ook)

**

**Pl

ease

not

e th

at b

oth

thes

e sy

stem

s in

clud

e Pe

er A

ssis

-ta

nce

and

Revi

ew c

ompo

nent

s an

d th

is c

hart

is o

nly

addr

ess-

ing

the

Peer

Ass

ista

nce

part

Evid

ence

(Hum

phre

y, e

t. al

, p. 2

2)

sugg

ests

that

CTs

tend

to c

ondu

ct

mor

e co

mpr

ehen

sive

ass

essm

ents

th

an p

rinci

pals

who

are

mor

e ty

pica

lly u

sing

a c

heck

list t

o id

entif

y is

sues

that

nee

d at

tent

ion.

CTs

typi

-ca

lly s

pend

mor

e tim

e w

ith te

ache

rs

durin

g PA

or P

AR

than

prin

cipa

ls o

r ad

min

istr

ator

s—an

d th

eir f

eedb

ack

is

ofte

n m

ore

supp

ortiv

e, le

ss th

reat

en-

ing,

and

bro

ader

in s

cope

.

• W

orks

bes

t whe

n de

sign

ed c

olla

bora

tivel

y w

ith

labo

r-m

anag

emen

t sta

keho

lder

s

•Tho

ugh

PA m

ay c

ost m

ore

initi

ally

, man

y sc

hool

di

stric

ts re

port

that

ove

r tim

e, P

A p

rogr

ams

redu

ce

turn

over

cos

ts, s

uppo

rt te

ache

r rec

ruitm

ent a

nd

rete

ntio

n, a

nd d

ecre

ase

the

time

and

expe

nse

as-

soci

ated

with

dis

mis

sing

car

eer t

each

ers

• C

an e

xpan

d ca

paci

ty fo

r for

mat

ive

feed

back

• C

an w

ork

flexi

bly

with

in a

bui

ldin

g or

dis

tric

t’s

uniq

ue s

ched

ules

and

nee

ds

• M

ay fe

el le

ss th

reat

enin

g th

an re

ceiv

ing

feed

back

fro

m p

rinci

pal

• Co

uld

have

mor

e fle

xibi

lity

to h

ave

CT w

ith

sim

ilar e

xper

ienc

es, t

hus

feed

back

wou

ld b

e m

ore

valid

and

relia

ble

• PA

are

exp

ensi

ve in

bot

h tim

e an

d m

oney

to im

plem

ent.

A s

choo

l dis

tric

t m

ust h

ire te

ache

rs to

repl

ace

the

CTs

in

the

clas

sroo

m. O

ther

cos

ts a

re th

e st

i-pe

nds

paid

to C

Ts a

nd g

ener

al a

dmin

is-

trat

ive

and

trai

ning

cos

ts a

ssoc

iate

d w

ith

the

prog

ram

• CT

job

resp

onsi

bilit

ies

are

com

preh

en-

sive

and

dem

andi

ng –

mus

t be

thor

ough

-ly

thou

ght t

hrou

gh a

nd te

nabl

e

• CT

wou

ld n

eed

high

leve

l of t

rain

ing

and

stan

dard

izat

ion

to m

ake

resu

lts c

ompa

-ra

ble,

relia

ble,

and

val

id

• CT

s w

ould

nee

d to

be

car

eful

ly c

hose

n th

roug

h a

high

ly th

ough

tful p

roce

ss

• Th

is s

yste

m m

ust h

ave

stak

ehol

der

buy-

in u

p fro

nt –

this

mea

ns ti

me

and

mon

ey to

trai

n an

d ed

ucat

e al

l sta

keho

ld-

ers

prio

r to

an a

sk o

f buy

-in

• Th

is s

yste

m c

ould

lead

to a

det

erio

ra-

tion

of re

latio

nshi

ps b

etw

een

teac

hers

if

it is

not

impl

emen

ted

with

abu

ndan

t tr

aini

ng, r

esou

rces

, tim

e an

d th

ough

tful-

ness

Mea

sure

:

Perc

eptio

n Su

rvey

s (s

tude

nt/p

aren

ts) (

FORM

ATIV

E O

NLY

as

onl

y lic

ense

d ad

min

istr

ator

s ca

n ev

alua

te te

ache

rs p

er

the

Teac

her S

tand

ards

and

Pra

ctic

es C

omm

issi

on (T

SPC)

, w

ho h

as b

een

give

n th

e au

thor

ity in

sta

tute

to li

cens

e an

d de

fine

the

role

s of

Adm

inis

trat

ors

(ORS

342

.121 a

nd O

RS

342.

140)

and

they

hav

e th

eref

ore

defin

ed li

cens

ed A

dmin

-is

trat

ors

as th

e on

ly  “

teac

hers

” al

low

ed to

eva

luat

e an

d su

perv

ise

(OA

R 58

4-00

5-00

05) w

hich

is fu

rthe

r defi

ned

and

rest

ricte

d by

sta

te b

arga

inin

g la

w (P

ublic

Em

ploy

ee

Colle

ctiv

e Ba

rgai

ning

Act

) ORS

243

.650

Des

crip

tion:

Stud

ent &

Par

ent S

urve

ys:

The

se q

uest

ionn

aire

s ge

nera

lly a

sk s

tude

nts

& p

aren

ts to

ra

te te

ache

rs o

n an

ext

ent s

cale

/lic

hert

sca

le (e

.g.,

from

1 to

5, w

here

1 =

very

effe

ctiv

e, a

nd 5

= n

ot a

t all

effec

tive)

re

gard

ing

vario

us a

spec

ts o

f tea

cher

s’ p

ract

ice

(e.g

., co

urse

co

nten

t, us

eful

ness

of f

eedb

ack,

sta

rtin

g cl

asse

s on

tim

e), a

s w

ell a

s ho

w m

uch

stud

ents

say

they

lear

ned

or th

e ex

tent

to

whi

ch th

ey w

ere

enga

ged.

The

y ve

ry o

ften

are

not u

sed

for

teac

her e

valu

atio

n at

the

pre-

colle

giat

e le

vel.

Use

d to

gat

her

stud

ent o

pini

ons

or ju

dgm

ents

abo

ut te

achi

ng p

ract

ice

as

part

of t

each

er e

valu

atio

n an

d to

pro

vide

info

rmat

ion

abou

t te

achi

ng a

s it

is p

erce

ived

by

stud

ents

.

Exam

ples

:

Stud

ent S

urve

ys:

• Ro

nald

Fer

guso

n, T

ripod

Pro

ject

Sur

veys

at H

arva

rd U

nive

r-si

ty

• G

allu

p St

uden

t Eng

agem

ent S

urve

ys

• G

eorg

ia p

lans

to u

se te

ache

r-fo

cuse

d st

uden

t sur

veys

as

part

of t

heir

teac

her e

valu

atio

n sy

stem

sta

rtin

g at

Gra

de 4

(a

ccor

ding

to th

e G

A R

TTT

appl

icat

ion)

Pare

nt S

urve

ys:

• Ro

nald

Fer

guso

n, T

ripod

Pro

ject

Sur

veys

at H

arva

rd U

nive

r-si

ty

• U

tah’

s Pa

y fo

r Per

form

ance

pilo

t pro

gram

use

s pa

rent

sa

tisfa

ctio

n su

rvey

s •

Geo

rgia

pla

ns to

use

par

ent s

urve

ys fo

r tea

cher

s of

stu

dent

s in

Gra

des

K–3

(acc

ordi

ng to

the

GA

RTT

T ap

plic

atio

n)

Stud

ent S

urve

ys:

Seve

ral s

tudi

es h

ave

show

n th

at

stud

ent r

atin

gs o

f tea

cher

s ca

n be

us

eful

in p

rovi

ding

info

rmat

ion

abou

t te

achi

ng; m

ay b

e as

val

id a

s ju

dg-

men

ts m

ade

by c

olle

ge s

tude

nts

and

othe

r gro

ups;

and

, in

som

e ca

ses,

m

ay c

orre

late

with

mea

sure

s of

st

uden

t ach

ieve

men

t. Va

lidity

is d

e-pe

nden

t on

the

inst

rum

ent u

sed

and

its a

dmin

istr

atio

n an

d is

gen

eral

ly

reco

mm

ende

d fo

r for

mat

ive

use

only

.

• Ro

wan

, B.,

Corr

enti,

R.,

& M

iller

, R.

J. (2

002)

. “W

hat l

arge

-sca

le, s

urve

y re

sear

ch te

lls u

s ab

out t

each

er

effec

ts o

n st

uden

t ach

ieve

men

t: In

sigh

ts fr

om th

e ‘P

rosp

ects

’ stu

dy

of e

lem

enta

ry s

choo

ls.”

Phila

delp

hia:

C

PRE

Publ

icat

ions

.

Mot

ivat

ed s

tere

otyp

ing

of w

omen

: “s

he’s

fine

if sh

e pr

aise

d m

e bu

t in

com

pete

nt if

she

crit

iciz

ed m

e” w

as

done

by

soci

al p

sych

olog

ists

Sin

clai

r &

Kun

da; F

ound

sam

e eff

ect f

or ra

ce

in a

noth

er s

tudy

Stud

ent S

urve

ys:

•Pro

vide

s pe

rspe

ctiv

e of

stu

dent

s w

ho h

ave

the

mos

t exp

erie

nce

with

teac

hers

•Can

pro

vide

form

ativ

e in

form

atio

n to

hel

p te

ach-

ers

impr

ove

prac

tice

in a

way

that

will

con

nect

w

ith s

tude

nts

• M

akes

use

of s

tude

nts,

who

may

be

as c

apab

le

as a

dult

rate

rs a

t pro

vidi

ng a

ccur

ate

ratin

gs

• Pr

ovid

es p

ersp

ectiv

e of

sch

ools

’ prim

ary

clie

nts

• C

an p

rovi

de in

form

atio

n on

how

to im

prov

e re

latio

nshi

ps w

ith s

tude

nts

• H

ave

been

val

idat

ed in

cer

tain

con

text

s

• C

an p

rovi

de im

port

ant f

orm

ativ

e in

form

atio

n on

te

ache

r pra

ctic

e

Pare

nt S

urve

ys:

• C

an p

rovi

de p

ersp

ectiv

e of

sch

ools

’ prim

ary

clie

nts

• C

an p

rovi

de in

form

atio

n on

how

to im

prov

e re

latio

nshi

ps w

ith s

tude

nts

• M

ay p

rovi

de in

form

atio

n th

at c

an b

e us

ed fo

r to

info

rm te

achi

ng p

ract

ice

Stud

ent S

urve

ys:

• St

uden

t rat

ings

hav

e no

t bee

n va

lidat

ed

for u

se in

sum

mat

ive

asse

ssm

ent a

nd

shou

ld n

ot b

e us

ed a

s a

sole

or p

rimar

y m

easu

re o

f tea

cher

eva

luat

ion

• St

uden

ts c

anno

t pro

vide

info

rmat

ion

on

aspe

cts

of te

achi

ng s

uch

as a

teac

her’s

co

nten

t kno

wle

dge,

cur

ricul

um fu

lfill-

men

t, an

d pr

ofes

sion

al a

ctiv

ities

• St

uden

ts a

re n

ot a

ble

to o

bser

ve m

uch

of w

hat g

oes

into

a te

ache

rs’ p

ract

ice

and

ther

efor

e m

ight

not

cap

ture

impo

rtan

t in

form

atio

n

• H

ave

not b

een

valid

ated

for s

umm

ativ

e de

cisi

ons

in P

K–12

cla

ssro

oms

• H

ow d

o ve

ry y

oung

chi

ldre

n pa

rtic

i-pa

te?

Read

ing

leve

ls?

Pare

nt S

urve

ys:

• Pa

rent

s no

t abl

e to

obs

erve

muc

h of

w

hat g

oes

into

a te

ache

rs’ p

ract

ice

and

ther

efor

e m

ight

not

cap

ture

impo

rtan

t in

form

atio

n ab

out a

teac

hers

’ cla

ssro

om

prac

tice

• H

ave

not b

een

valid

ated

for s

umm

ativ

e de

cisi

ons

• D

o no

t inc

lude

relia

ble

info

rmat

ion

on

stud

ent l

earn

ing

grow

th

• Pa

rtic

ipat

ion

is o

ften

low

, the

refo

re

resu

lting

info

rmat

ion

that

may

not

be

valid

or r

elia

ble

• La

ngua

ge b

arrie

rs, l

itera

cy ra

tes,

or

othe

r fac

tors

may

pre

vent

par

ents

from

pa

rtic

ipat

ing

in s

urve

ys.

Page 32: OEA's Teacher Evaluation & Support System Guidebook

3232

Page 33: OEA's Teacher Evaluation & Support System Guidebook

33

Student Learning and Growth: Evidence of teachers’ contribution to student learning and growth.

The Oregon Framework requires at least two student learning goals and the identification of strategies and measures that will be used to determine goal attainment

(see table to the right). Teachers also specify what evidence will be provided to document progress on each goal:

1. Teachers who are responsible for student learning in tested subjects and grades (i.e., ELA and mathematics in grades 3-8, 11) will use state assessments as one measure (category 1) and will also select one or more additional measures from category 2 or 3 that provide additional evidence of students’ growth and proficiency/mastery of the standards, and evidence of deeper learning and 21st century skills.

2. Teachers in non-tested (state test) subjects and grades will use measures that are valid representations of student learning standards from at least two of the following three categories, based on what is most appropriate for the cur-riculum and students they teach. [OEA recognizes this as a state requirement, but this may be a complex task as many subjects/content areas do not have measures outside of category 3.]

Student Learning Goals V.

Student Learning and Growth as a Measure of Teaching Effectiveness

Student Learning and Growth Goal-Setting Process • Teachers review baseline data and create goals that measure

the learning of all students. Goals span a school year or com-plete course of study. OEA recommends you write growth goals on skill acquisition or content as demonstrated by evi-dence from multiple student learning measures, and do not write student learning goals about growth in the number of students “passing” particular assessments or student learn-ing measures (i.e., do not write goals that read “10% more students will pass the OAKS reading test”, or “15% more stu-dents will receive all 4s or higher on their expository writing samples”). See the next page for examples of SMART goals.

• Teachers collaborate with supervisor/evaluator to establish student learning goals. In addition, teachers may collaborate to establish student learning goals for their grade levels, departments, or curricular teams.

• Teachers will establish at least two student learning goals and identify strategies and measures that will be used to determine goal attainment. They also specify what evidence will be provided to document progress on each goal. OEA recommends that each student learning goal contain mul-tiple measures of student learning and to avoid a workload issue by using measures of student learning already in use in your classroom. For example, in a student learning goal created by a teacher in a “tested grade or subject,” a teacher would create a goal around improving or showing growth in a particular reading skill then would include a variety of stu-dent learning measures to show evidence of progress toward this goal, i.e., OAKS, DIBELS, a curriculum-based assess-ment, and a sample of students’ work/portfolios.

• Teachers complete goal setting in collaboration with their supervisor/evaluator. During the collaborative planning process, the teacher and supervisor/evaluator ensure that quality goal setting occurs through a discussion of the rigor and rationale of each goal, appropriate research-based strategies, quality of evidence and standards addressed. The SMART goal process is used in the development of student growth goals (SMART = Specific and Strategic; Measurable; Action oriented; Rigorous, Realistic, and Results-focused; Timed and Tracked).

• Teachers meet with supervisor/evaluator to discuss prog-ress for each goal mid-year and at the end of the year. Goals remain the same throughout the year, but strategies for at-taining goals may be revised.

• Teachers, along with their supervisor/evaluator, reflect on the results and determine implications for future profes-sional growth planning.

33

Page 34: OEA's Teacher Evaluation & Support System Guidebook

34

1. SMART Goal Format During the 20__-__ school year, __% of students will improve their ____________skills by __% as measured by evidence from: __measure of student learning___, __measure of student learning___, and __measure of student learning___. **Note – can be more than three measures**

2. SMART Goal = Writing During the 2012-13 school year, 80% of students will improve their expository writing skills by 10% as measured by evidence from: assessment for chapter(s) 7-9 on expository writing skills, student writing samples graded on district ru-bric, and curriculum pre-/post-test.

3. SMART Goal = Math During the 2012-13 school year, 90% of students will improve their math prob-lem-solving skills by 10% as measured by evidence from 2012-13 math problem solving strand data of OAKS, easy CBMs, curriculum assessment for chapter(s) 2-3, and student work samples.

Category Types of Measures of Student Learning Examples include, but are not limited to:

1 State or national standardized testsOregon Assessment of Knowledge and Skills (OAKSSMARTER: Balanced (when adopted), English Language Profi-ciency Assessment (ELPA), Extended Assessments

2 Common national, international, regional, district-developed measures

ACT, PLAN, EXPLORE, AP, IB, DIBELS, C-PAS, other national mea-sures; or common assessments approved by the district or state as valid, reliable and able to be scored comparably across schools or classrooms

3 Classroom-based or school-wide measures

Student performances, portfolios, products, projects, work samples, tests

Student Learning Goals V.

EXAMPLES OF SMART GOALS:

Page 35: OEA's Teacher Evaluation & Support System Guidebook

35

I. SCHOOL LEVEL — INFORM INSTRUCTION AND PROGRAMS

Use of Measure(s) of Student Learning Value of UseIncorporating multiple valid measures of student learning to be used in a whole school Formative Evaluation process, designed to inform and improve practice

Low Stakes, High Efficacy

One measure of student learning (ex: just using OAKS/ statewide assessment scores) as sole measure of student learning to be used as a whole school For-mative Evaluation process, designed to inform and improve practice

Low Stakes, Low Efficacy

One measure of student learning to be used as a “trigger” to examine a school’s performance more closely, but not to be used as part of a final or Summative Evaluation, or used for high-stakes decision making

Moderate Stakes, Low Efficacy

II. INDIVIDUAL TEACHER LEVEL — INFORM INSTRUCTION

Use of Measure(s) of Student Learning Value of UseIncorporating multiple valid measures of student learning to be used in an individual teacher Formative Evaluation process, designed to inform and improve practice

Low Stakes, High Efficacy

One measure of student learning (ex: just using OAKS/statewide assess-ment scores) as sole measure of student learning to be used in an individual teacher Formative Evaluation process, designed to inform and improve prac-tice

Moderate Stakes, Low Efficacy

One measure of student learning to be used as a “trigger” to examine a teacher’s performance more closely, but not to be used as part of a final or Summative Evaluation, or used for high-stakes decision making such as hiring, firing, or compensation, etc.

Moderate -High Stakes, Low Efficacy

III. INDIVIDUAL TEACHER LEVEL — EVALUATE PERFORMANCECareful consideration should be given to the “weight” of measures of student learning against other measures

Use of Measure(s) of Student Learning Value of UseIncorporating multiple valid measures of student learning and student learn-ing then becoming one of multiple valid measures of teaching effectiveness – multiple measures within multiple measures.

CAUTION: Potential for Moderate to High Stakes, Moderate Efficacy• Most measures of student learning do not have sub-

stantial field research validating them as measures of teaching effectiveness

• Moderate efficacy due to other (multiple) measures besides student learning measures

One measure of student learning to be used as a significant percentage of a Summative Evaluation or other high-stakes decision.

EXTREME CAUTION: High Stakes, Low Efficacy

One measure of student learning to be used as a sole measure for a Sum-mative Evaluation system, or sole measure for high-stakes decisions such as hiring, firing, compensation, etc.

STOP: High Stakes, Low Efficacy

OREGON EDUCATION ASSOCIATION’S RECOMMENDED USE OFSTUDENT DATA IN TEACHER EVALUATION AND SUPPORT SYSTEMS

Using Measures of Student Learning VI.

Page 36: OEA's Teacher Evaluation & Support System Guidebook

36

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

ipti

on /

Exa

mpl

esRe

sear

chSt

reng

ths

Cauti

ons

Clas

sroo

m-le

vel:

Stud

ent g

row

th a

nd/o

r st

uden

t lea

rnin

g ta

rget

s de

velo

ped

colla

bora

tive

ly b

etw

een

teac

her

and

prin

cipa

l for

the

aca

dem

ic y

ear

Mea

sure

:

Stud

ent G

row

th O

bjec

tives

or G

oal-D

riven

Po

rtfo

lio P

lans

Des

crip

tion:

Thes

e ev

alua

tion

tool

s off

er te

ache

rs th

e op

port

unity

to s

et th

eir o

wn

high

but

feas

ible

ob

ject

ives

for t

heir

stud

ents

’ gro

wth

in c

ol-

labo

ratio

n w

ith th

eir p

rinci

pal a

nd/o

r oth

er

colle

ague

s. T

he a

sses

smen

ts te

ache

rs u

se m

ay

be c

omm

on o

r sta

ndar

dize

d ex

ams

or te

ache

r-de

velo

ped

asse

ssm

ents

. Som

e to

ols

requ

ire

teac

hers

to s

peci

fy th

e pr

ofes

sion

al d

evel

op-

men

t the

y w

ill p

artic

ipat

e in

to e

nsur

e th

eir

stud

ents

ach

ieve

thei

r gro

wth

obj

ectiv

es.

Exam

ples

:

• St

uden

t Lea

rnin

g O

bjec

tives

- Aus

tin In

depe

n-de

nt S

choo

l Dis

tric

t

• St

uden

t Gro

wth

Obj

ectiv

es- D

enve

r Pub

lic

Scho

ols

• N

ew M

exic

o Pr

ofes

sion

al D

evel

opm

ent P

lans

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

”•

Can

be

used

to g

auge

teac

hers

’ con

trib

utio

ns

to o

utco

mes

in u

ntes

ted

subj

ects

(e.g

., so

cial

st

udie

s, b

iolo

gy, m

usic

)

• C

an fo

cus

teac

hers

’ pra

ctic

e on

ach

ievi

ng

part

icul

ar o

utco

mes

bas

ed o

n th

eir a

naly

ses

of

stud

ent l

earn

ing

need

s

• C

an p

rom

ote

colla

bora

tion

amon

g te

ache

rs

and

inst

ruct

iona

l lea

ders

• C

an in

cent

iviz

e te

ache

rs to

eng

age

in p

rofe

s-si

onal

lear

ning

opp

ortu

nitie

s th

at w

ill h

elp

them

ac

hiev

e go

als

• Sa

fegu

ards

mus

t be

in p

lace

to e

nsur

e th

at

goal

s se

t are

feas

ible

yet

als

o hi

gh a

nd ri

goro

us

• Co

mpa

rabi

lity

amon

g te

ache

rs m

ay b

e pr

ob-

lem

atic

Mea

sure

:

Clas

sroo

m A

sses

smen

ts

Des

crip

tion:

Ass

essm

ents

inte

grat

ed th

roug

hout

dis

tric

t-w

ide

curr

icul

a, o

r ind

ivid

ually

dev

elop

ed a

t the

te

ache

r, bu

ildin

g, o

r dis

tric

t lev

el to

mea

sure

st

uden

t pro

gres

s th

roug

h cu

rric

ula

and

com

-pa

red

to s

tand

ards

Exam

ples

:

NC

LB W

aive

r exa

mpl

es

NW

ESD

has

don

e so

me

rese

arch

on

thes

e m

easu

res

• M

ultip

le ty

pes

• C

an b

e qu

ick

or le

ngth

y to

adm

inis

ter

• A

ltern

ativ

e to

larg

e-sc

ale

test

s

• M

ay h

ave

relia

bilit

y is

sues

• Ti

me

cons

umin

g fo

r ind

ivid

ual t

each

ers

to

crea

te

• Co

mpa

rabi

lity

amon

g te

ache

rs m

ay b

e pr

ob-

lem

atic

Wha

t You

're Lo

okin

g At

...

The

follo

win

g ch

art p

rovi

des

you

with

sev

eral

exa

mpl

es o

f m

easu

res

of s

tude

nt le

arni

ng. F

or e

ach

mea

sure

, the

re is

a

desc

riptio

n w

ith e

xam

ples

, cur

rent

rese

arch

, and

a li

st o

f the

m

easu

re’s

stre

ngth

s as

wel

l as

caut

ions

for u

se.

Page 37: OEA's Teacher Evaluation & Support System Guidebook

37

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

ipti

on /

Exa

mpl

esRe

sear

chSt

reng

ths

Cauti

ons

Clas

sroo

m-le

vel:

Stud

ent g

row

th a

nd/o

r st

uden

t lea

rnin

g ta

rget

s de

velo

ped

colla

bora

tive

ly b

etw

een

teac

her

and

prin

cipa

l for

the

aca

dem

ic y

ear

Mea

sure

:

Stud

ent G

row

th O

bjec

tives

or G

oal-D

riven

Po

rtfo

lio P

lans

Des

crip

tion:

Thes

e ev

alua

tion

tool

s off

er te

ache

rs th

e op

port

unity

to s

et th

eir o

wn

high

but

feas

ible

ob

ject

ives

for t

heir

stud

ents

’ gro

wth

in c

ol-

labo

ratio

n w

ith th

eir p

rinci

pal a

nd/o

r oth

er

colle

ague

s. T

he a

sses

smen

ts te

ache

rs u

se m

ay

be c

omm

on o

r sta

ndar

dize

d ex

ams

or te

ache

r-de

velo

ped

asse

ssm

ents

. Som

e to

ols

requ

ire

teac

hers

to s

peci

fy th

e pr

ofes

sion

al d

evel

op-

men

t the

y w

ill p

artic

ipat

e in

to e

nsur

e th

eir

stud

ents

ach

ieve

thei

r gro

wth

obj

ectiv

es.

Exam

ples

:

• St

uden

t Lea

rnin

g O

bjec

tives

- Aus

tin In

depe

n-de

nt S

choo

l Dis

tric

t

• St

uden

t Gro

wth

Obj

ectiv

es- D

enve

r Pub

lic

Scho

ols

• N

ew M

exic

o Pr

ofes

sion

al D

evel

opm

ent P

lans

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

”•

Can

be

used

to g

auge

teac

hers

’ con

trib

utio

ns

to o

utco

mes

in u

ntes

ted

subj

ects

(e.g

., so

cial

st

udie

s, b

iolo

gy, m

usic

)

• C

an fo

cus

teac

hers

’ pra

ctic

e on

ach

ievi

ng

part

icul

ar o

utco

mes

bas

ed o

n th

eir a

naly

ses

of

stud

ent l

earn

ing

need

s

• C

an p

rom

ote

colla

bora

tion

amon

g te

ache

rs

and

inst

ruct

iona

l lea

ders

• C

an in

cent

iviz

e te

ache

rs to

eng

age

in p

rofe

s-si

onal

lear

ning

opp

ortu

nitie

s th

at w

ill h

elp

them

ac

hiev

e go

als

• Sa

fegu

ards

mus

t be

in p

lace

to e

nsur

e th

at

goal

s se

t are

feas

ible

yet

als

o hi

gh a

nd ri

goro

us

• Co

mpa

rabi

lity

amon

g te

ache

rs m

ay b

e pr

ob-

lem

atic

Mea

sure

:

Clas

sroo

m A

sses

smen

ts

Des

crip

tion:

Ass

essm

ents

inte

grat

ed th

roug

hout

dis

tric

t-w

ide

curr

icul

a, o

r ind

ivid

ually

dev

elop

ed a

t the

te

ache

r, bu

ildin

g, o

r dis

tric

t lev

el to

mea

sure

st

uden

t pro

gres

s th

roug

h cu

rric

ula

and

com

-pa

red

to s

tand

ards

Exam

ples

:

NC

LB W

aive

r exa

mpl

es

NW

ESD

has

don

e so

me

rese

arch

on

thes

e m

easu

res

• M

ultip

le ty

pes

• C

an b

e qu

ick

or le

ngth

y to

adm

inis

ter

• A

ltern

ativ

e to

larg

e-sc

ale

test

s

• M

ay h

ave

relia

bilit

y is

sues

• Ti

me

cons

umin

g fo

r ind

ivid

ual t

each

ers

to

crea

te

• Co

mpa

rabi

lity

amon

g te

ache

rs m

ay b

e pr

ob-

lem

atic

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

ipti

on /

Exa

mpl

esRe

sear

chSt

reng

ths

Cauti

ons

Scho

ol-w

ide:

(bui

ldin

g le

vel)

Stud

ent

grow

th a

nd/o

r ac

hiev

emen

t as

dete

rmin

ed b

y ap

prov

ed s

tate

wid

e as

sess

men

t sys

tem

(i.e

., O

AKS

, SM

ART

ER)

Mea

sure

:

Stan

dard

ized

Tes

ts:

Des

crip

tion:

Incl

udes

, but

not

lim

ited

to, c

ompa

rabl

e, n

orm

-re

fere

nced

sum

mat

ive

or fo

rmat

ive

asse

ss-

men

ts s

core

d an

d ad

min

iste

red

in a

con

sist

ent

man

ner.

Exam

ples

:

Nat

iona

l, st

ate

or d

istr

ict a

sses

smen

ts. (

See

Ore

gon

Fram

ewor

k fo

r Tea

cher

& A

dmin

istr

ator

Ev

alua

tion

& S

uppo

rt S

yste

m o

n C

D-R

OM

)

•htt

p://

ww

w.te

achi

ngqu

ality

.org

/leg

acy/

DQ

C_

Qua

rter

lyIs

sueB

rief.p

df

• A

FT p

g 31

-34

inte

rest

ing

mod

el fr

om G

eorg

ia

http

://w

ww

.isbe

.net

/PEA

C/p

df/m

ultip

le_m

ea-

sure

s_in

tro_

0411

.pdf

Reco

mm

enda

tions

for b

uild

ing

data

sys

tem

s ht

tp://

ww

w.d

ataq

ualit

ycam

paig

n.or

g/fil

es/

DQ

C_T

SDL_

7-27

.pdf

• Q

uant

itativ

e

• Co

nsis

tent

• Le

ss s

ubje

ctiv

e

•Sta

rts

to m

easu

re im

pact

of i

nstr

uctio

n on

st

uden

t per

form

ance

• Te

st s

core

gai

ns a

re m

ore

than

an

indi

vidu

al

teac

hers

effo

rt

• Te

ache

rs ra

tings

bas

ed o

n Va

lue-

Add

ed m

od-

els

are

unst

able

with

larg

e m

argi

ns o

f err

or a

nd

dram

atic

sw

ings

• M

ay p

rom

ote

teac

hing

to th

e te

st a

nd le

ave

less

tim

e fo

r tea

chin

g hi

gher

-leve

l thi

nkin

g sk

ills

• C

an p

unis

h te

ache

rs w

ho w

ork

with

the

mos

t cha

lleng

ing

or h

isto

rical

ly u

nder

serv

ed

popu

latio

ns o

f stu

dent

s, a

s w

ell a

s th

ose

who

w

ork

with

hig

h-le

vel s

tude

nts

(floo

r and

cei

ling

issu

es)

• N

o co

untr

y in

the

wor

ld u

ses

annu

al te

st

scor

e ga

ins

to e

valu

ate

thei

r tea

cher

s, in

clud

ing

the

high

est-

leve

l per

form

ing

coun

trie

s

• D

oesn

’t lin

k sp

ecifi

c te

ache

r pra

ctic

es/b

ehav

-io

r to

impr

oved

stu

dent

per

form

ance

• Th

e AY

P m

odel

com

pare

s pe

rform

ance

of

diffe

rent

coh

orts

of s

tude

nts

• H

ow d

o yo

u m

aint

ain

an e

valu

atio

n sy

stem

w

here

onl

y ce

rtai

n te

ache

rs h

ave

acce

ss to

st

anda

rdiz

ed te

sts?

• M

ost t

ests

not

sca

led

acro

ss y

ears

.

• Ev

en s

ophi

stic

ated

VA

M m

odel

s ca

nnot

ac-

coun

t for

team

teac

hing

and

oth

er e

ffect

s.

Page 38: OEA's Teacher Evaluation & Support System Guidebook

38

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

ipti

on /

Exa

mpl

esRe

sear

chSt

reng

ths

Cauti

ons

Scho

ol-w

ide:

(bui

ldin

g le

vel)

Stud

ent

grow

th a

nd/o

r ac

hiev

emen

t as

dete

rmin

ed b

y ap

prov

ed s

tate

wid

e as

sess

men

t sys

tem

(i.e

., O

AKS

, SM

ART

ER)

Mea

sure

:

Mea

sure

s of

Stu

dent

Eng

agem

ent o

r Edu

ca-

tiona

l Att

ainm

ent

Des

crip

tion:

Thes

e ca

n in

clud

e cl

assr

oom

obs

erva

tions

, an

d/or

sel

f-re

port

s of

stu

dent

eng

agem

ent,

as

wel

l as

mea

surin

g A

dvan

ced

Plac

emen

t cou

rse

part

icip

atio

n ra

tes,

gra

duat

ion

rate

s, d

ropo

ut

rate

s, a

nd s

tude

nt a

bsen

teei

sm.

Exam

ples

:

• U

mat

illa

Scho

ol D

istr

ict,

Ore

gon

(see

CD

-RO

M)

• G

eorg

ia, w

here

acc

ordi

ng to

the

Geo

rgia

Rac

e to

the

Top

appl

icat

ion,

“pla

ns to

inve

st in

the

deve

lopm

ent,

test

ing,

and

eva

luat

ion

of a

ltern

a-tiv

e qu

antit

ativ

e m

easu

res

to a

sses

s st

uden

t en

gage

men

t and

stu

dent

ach

ieve

men

t”. T

he

Mar

ylan

d RT

T ap

plic

atio

n su

gges

ts th

at lo

cal

scho

ol s

yste

ms

can

“pro

pose

alte

rnat

ive

prio

ri-tie

s fo

r ann

ually

mea

surin

g st

uden

t gro

wth

and

le

arni

ng, s

uch

as—

at th

e hi

gh-s

choo

l lev

el—

gain

s in

Adv

ance

d Pl

acem

ent p

artic

ipat

ion

and

exam

per

form

ance

or d

ecre

ases

in th

e dr

opou

t ra

te”.

The

Mea

sure

s of

Effe

ctiv

e Te

achi

ng S

tudy

fu

nded

by

the

Gat

es F

ound

atio

n is

cur

rent

ly

exam

inin

g a

mea

sure

of s

tude

nt e

ngag

emen

t as

a va

lid p

redi

ctor

of s

tude

nt g

row

th. T

he T

ripod

Su

rvey

was

use

d in

this

stu

dy to

mea

sure

stu

-de

nt e

ngag

emen

t.

• C

an a

sses

s an

d in

cent

iviz

e ot

her i

mpo

rtan

t te

achi

ng o

utco

mes

• C

an p

rovi

de s

ome

form

ativ

e an

d su

mm

ativ

e in

form

atio

n on

the

effec

ts o

f tea

cher

s’ p

ract

ice

• D

ifficu

lt to

att

ribut

e in

divi

dual

teac

her c

ontr

i-bu

tion

to s

uch

outc

omes

• H

ave

not b

een

wid

ely

test

ed in

the

field

to

dete

rmin

e va

lidity

, rel

iabi

lity,

etc

.

Page 39: OEA's Teacher Evaluation & Support System Guidebook

39

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

ipti

on /

Exa

mpl

esRe

sear

chSt

reng

ths

Cauti

ons

Scho

ol-w

ide:

(bui

ldin

g le

vel)

Stud

ent

grow

th a

nd/o

r ac

hiev

emen

t as

dete

rmin

ed b

y ap

prov

ed s

tate

wid

e as

sess

men

t sys

tem

(i.e

., O

AKS

, SM

ART

ER)

Mea

sure

:

Mea

sure

s of

Stu

dent

Eng

agem

ent o

r Edu

ca-

tiona

l Att

ainm

ent

Des

crip

tion:

Thes

e ca

n in

clud

e cl

assr

oom

obs

erva

tions

, an

d/or

sel

f-re

port

s of

stu

dent

eng

agem

ent,

as

wel

l as

mea

surin

g A

dvan

ced

Plac

emen

t cou

rse

part

icip

atio

n ra

tes,

gra

duat

ion

rate

s, d

ropo

ut

rate

s, a

nd s

tude

nt a

bsen

teei

sm.

Exam

ples

:

• U

mat

illa

Scho

ol D

istr

ict,

Ore

gon

(see

CD

-RO

M)

• G

eorg

ia, w

here

acc

ordi

ng to

the

Geo

rgia

Rac

e to

the

Top

appl

icat

ion,

“pla

ns to

inve

st in

the

deve

lopm

ent,

test

ing,

and

eva

luat

ion

of a

ltern

a-tiv

e qu

antit

ativ

e m

easu

res

to a

sses

s st

uden

t en

gage

men

t and

stu

dent

ach

ieve

men

t”. T

he

Mar

ylan

d RT

T ap

plic

atio

n su

gges

ts th

at lo

cal

scho

ol s

yste

ms

can

“pro

pose

alte

rnat

ive

prio

ri-tie

s fo

r ann

ually

mea

surin

g st

uden

t gro

wth

and

le

arni

ng, s

uch

as—

at th

e hi

gh-s

choo

l lev

el—

gain

s in

Adv

ance

d Pl

acem

ent p

artic

ipat

ion

and

exam

per

form

ance

or d

ecre

ases

in th

e dr

opou

t ra

te”.

The

Mea

sure

s of

Effe

ctiv

e Te

achi

ng S

tudy

fu

nded

by

the

Gat

es F

ound

atio

n is

cur

rent

ly

exam

inin

g a

mea

sure

of s

tude

nt e

ngag

emen

t as

a va

lid p

redi

ctor

of s

tude

nt g

row

th. T

he T

ripod

Su

rvey

was

use

d in

this

stu

dy to

mea

sure

stu

-de

nt e

ngag

emen

t.

• C

an a

sses

s an

d in

cent

iviz

e ot

her i

mpo

rtan

t te

achi

ng o

utco

mes

• C

an p

rovi

de s

ome

form

ativ

e an

d su

mm

ativ

e in

form

atio

n on

the

effec

ts o

f tea

cher

s’ p

ract

ice

• D

ifficu

lt to

att

ribut

e in

divi

dual

teac

her c

ontr

i-bu

tion

to s

uch

outc

omes

• H

ave

not b

een

wid

ely

test

ed in

the

field

to

dete

rmin

e va

lidity

, rel

iabi

lity,

etc

.

Mea

sure

:

Mea

sure

s of

Stu

dent

Wor

k Th

at S

how

Ev

iden

ce o

f Gro

wth

Des

crip

tion

:

Thes

e m

ay ta

ke th

e fo

rm o

f alte

rnat

ive

asse

ssm

ents

of s

tude

nt le

arni

ng, f

or

exam

ple,

writ

ing

sam

ples

, por

tfol

ios

of

stud

ent w

ork,

stu

dent

ora

l pre

sent

a-tio

ns, c

apst

one

proj

ects

, and

the

like.

Exam

ples

:

Mas

sach

uset

ts w

ill w

ork

to d

evel

op

stud

ent p

erfo

rman

ce a

sses

smen

ts, a

nd

teac

hers

and

teac

her t

eam

s w

ill re

quire

tr

aini

ng o

n ho

w to

gat

her s

tude

nt w

ork

that

dem

onst

rate

s in

divi

dual

stu

dent

le

arni

ng, a

s w

ell a

s ty

pica

l stu

dent

le

arni

ng in

a c

lass

. The

sta

te a

lso

will

de

velo

p di

stric

t-ba

sed

asse

ssm

ents

th

at a

re c

ompa

rabl

e ac

ross

sub

ject

s an

d gr

ades

. Fin

ally

, the

sta

te w

ill in

clud

e st

uden

t wor

k sa

mpl

es a

s ev

iden

ce o

f st

uden

t lea

rnin

g.

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

”•

Can

be

mor

e au

then

tic a

sses

smen

ts o

f st

uden

t lea

rnin

g th

an s

tand

ardi

zed

test

s an

d th

eref

ore

mor

e va

lid a

sses

smen

ts o

f te

ache

rs’ c

ontr

ibut

ions

to s

tude

nt le

arn-

ing

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

form

atio

n ab

out a

teac

her’s

pr

actic

e

• V

alid

ity a

nd re

liabi

lity

not w

ell e

stab

-lis

hed

• V

ery

diffi

cult

to s

tand

ardi

ze a

nd

ther

efor

e di

fficu

lt to

est

ablis

h co

mpa

-ra

bilit

y an

d re

liabi

lity

• Ps

ycho

met

ric p

rope

rtie

s of

suc

h as

-se

ssm

ents

are

not

wel

l und

erst

ood

Mea

sure

:

Gai

n Sc

ore

Mod

els

or P

re-T

est/

Post

-tes

t M

etho

ds*

Des

crip

tion

:

Gai

n Sc

ore

Mod

els

mea

sure

the

diffe

r-en

ce b

etw

een

an e

arlie

r and

a la

ter t

est

scor

e, s

o un

like

Val

ue-A

dded

Mod

els

(VA

Ms)

, the

y on

ly re

quire

two

test

sc

ores

. Thi

s ap

proa

ch a

lso

relie

s on

ve

rtic

ally

equ

ated

ass

essm

ents

.

Exam

ples

:

Hill

sbor

ough

Cou

nty,

FL;

Eag

le C

ount

y,

CO

; and

MA

, MD

, and

NY

am

ong

othe

r Ra

ce to

the

Top

win

ners

will

dev

elop

pr

e- a

nd p

ost-

test

mea

sure

s of

stu

dent

le

arni

ng fo

r tea

cher

s in

sub

ject

s no

t te

sted

by

the

stat

e as

sess

men

t sys

tem

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

”•

Are

pre

fera

ble

to “

stat

us”

mod

els

beca

use

they

indi

cate

cha

nge

in s

tude

nt

lear

ning

ove

r tim

e

• C

an p

rovi

de in

form

atio

n on

som

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

out

com

es

• A

llow

s fo

r com

paris

ons

betw

een

teac

h-er

s

• C

an p

rovi

de s

ome

form

ativ

e an

d su

m-

mat

ive

info

rmat

ion

on th

e eff

ects

of

teac

hers

’ pra

ctic

e

• C

anno

t ade

quat

ely

cont

rol f

or s

tu-

dent

s’ b

ackg

roun

d ch

arac

teris

tics

• C

an b

e pr

one

to e

rror

if te

sts

are

too

easy

or t

oo d

ifficu

lt

• C

an le

ad to

fals

e at

trib

utio

ns o

f va

lue—

that

is, i

n te

am- o

r co-

teac

hing

si

tuat

ions

, can

not s

epar

ate

out c

ontr

i-bu

tions

of o

ne te

ache

r or a

noth

er

• G

row

th m

odel

s ca

n en

cour

age

non-

educ

ativ

e te

st p

rep,

che

atin

g, e

tc.

• V

alid

ity a

nd re

liabi

lity

can

only

be

achi

eved

if th

e ad

min

istr

atio

n of

the

test

s an

d in

terp

reta

tion

of th

e sc

ores

ar

e co

nsis

tent

acr

oss

all c

lass

room

s

• G

row

th m

odel

s ca

n di

s-in

cent

iviz

e co

llabo

ratio

n de

pend

ing

on h

ow th

ey

are

used

Page 40: OEA's Teacher Evaluation & Support System Guidebook

40

Mea

sure

:

Val

ue-A

dded

Mod

els*

Des

crip

tion

:

Val

ue-A

dded

Mod

els

(VA

Ms)

are

cur

-re

ntly

bei

ng u

sed

to d

eter

min

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

s’ te

st

scor

e ga

ins

and

also

are

bei

ng u

sed

as a

rese

arch

tool

(e.g

., de

term

inin

g th

e di

strib

utio

n of

“effe

ctiv

e” te

ache

rs

by s

tude

nt o

r sch

ool c

hara

cter

istic

s).

VAM

’s m

easu

re th

e ga

ins

that

stu

dent

s m

ake

and

adju

st th

ose

gain

s fo

r stu

dent

, te

ache

r, or

sch

ool c

hara

cter

istic

s. T

he

gain

s ar

e in

terp

rete

d as

the

“val

ue”

that

a

teac

her a

dds

beca

use

the

gain

s ar

e pr

esum

ed to

be

net o

f all

othe

r infl

uenc

-es

. Thi

s pr

esum

ptio

n is

muc

h de

bate

d.

Stro

nger

VA

Ms

use

wel

l-de

sign

ed v

erti-

cally

equ

ated

sta

ndar

dize

d ac

hiev

emen

t te

sts

that

mea

sure

rele

vant

con

cept

s an

d te

st s

tude

nts

at le

ast t

hree

tim

es.

Som

e te

sts

are

vert

ical

ly e

quat

ed, w

hich

m

eans

that

a g

iven

sco

re o

n th

e fo

urth

-gr

ade

vers

ion

of a

test

repr

esen

ts th

e sa

me

leve

l of p

erfo

rman

ce a

s th

at s

ame

scor

e on

the

fifth

-gra

de v

ersi

on o

f the

te

st. O

AKS

is n

ot a

ver

tical

ly e

quat

ed

asse

ssm

ent.

Exam

ples

:

• C

LASS

Pro

ject

TIF

Gra

nt re

cipi

ents

, su

ch a

s Sa

lem

-Kei

zer S

choo

l Dis

tric

t, Le

bano

n Sc

hool

Dis

tric

t; al

so M

cMin

-nv

ille

Scho

ol D

istr

ict

• Te

nnes

see

Val

ue-A

dded

Ass

essm

ent

Syst

em (

TVA

AS)

• O

hio’

s Ed

ucat

ion

Val

ue-A

dded

As-

sess

men

t Sys

tem

• D

alla

s V

alue

-Add

ed A

ccou

ntab

ility

Sy

stem

(D

VAA

S)

• Re

sear

cher

s ha

ve w

arne

d ag

ains

t usi

ng v

alue

-add

ed

estim

ates

for h

igh

stak

es p

urpo

ses,

incl

udin

g hi

ring,

fir-

ing,

com

pens

atio

n, a

nd s

umm

ativ

e ev

alua

tions

. (Ba

ker,

et a

l., 2

010

; Bra

un, 2

00

5; B

raun

, Chu

dow

sky

& K

onei

g,

2010

; Cor

cora

n, 2

020

; Kup

erm

intz

, 20

03)

• La

ck o

f con

sens

us in

the

rese

arch

com

mun

ity a

bout

w

heth

er V

AM

s ca

n ac

cura

tely

isol

ate

the

effec

ts o

f a

sing

le te

ache

r, es

peci

ally

ove

r an

exte

nded

per

iod.

(G

olds

chm

idt,

et a

l., 2

00

5; K

oede

l & B

etts

, 20

09;

Kan

e &

Sta

iger

, 20

08)

• Te

ache

r inp

ut m

ay b

e an

impo

rtan

t infl

uenc

e on

ac

hiev

emen

t, bu

t it i

s no

t the

onl

y in

fluen

ce.

(Bra

un,

200

5)

• VA

M n

umbe

rs d

o no

t offe

r for

mat

ive

feed

back

abo

ut

way

s to

impr

ove

inst

ruct

ion

or c

lass

room

effe

ctiv

enes

s.

(Mill

er, 2

00

9; R

owan

, Cor

rent

i, &

Mill

er, 2

00

2)

• VA

Ms

fail

to a

ccur

atel

y re

flect

the

limita

tions

of p

ar-

ticul

ar te

sts

both

for m

easu

ring

the

full

rang

e of

des

ired

know

ledg

e an

d sk

ills

and

for m

easu

ring

lear

ning

gai

ns.

(New

ton,

Dar

ling-

Ham

mon

d, H

aert

el &

Tho

mas

, 20

10)

• St

uden

ts a

re n

ot n

eces

saril

y as

sign

ed to

cla

ssro

oms

and

teac

hers

on

a ra

ndom

bas

is. T

his

has

maj

or ra

mi-

ficat

ions

on

man

y VA

Ms.

Thi

s w

ill s

ever

ely

impa

ct th

e va

lidity

and

relia

bilit

y of

a v

alue

-add

ed s

core

. Acc

urat

e VA

Ms

requ

ire c

ompa

rabi

lity

acro

ss d

emog

raph

ics

of a

n in

divi

dual

teac

her’s

cla

ssro

om o

r sch

ool s

ettin

g. (

Har

-ris

, 20

09)

• VA

Ms

are

expe

nsiv

e to

ado

pt a

nd m

aint

ain.

Mos

t sc

hool

dis

tric

ts c

anno

t affo

rd to

upk

eep

the

rigor

ous

data

requ

irem

ents

or t

he d

egre

e of

hum

an re

sour

ces

and

psyc

hom

etric

exp

ertis

e re

quire

d. (G

olds

chm

idt,

200

5; H

arris

, 20

09)

• Li

ttle

is k

now

n ab

out t

he v

alid

ity o

f val

ue-a

dded

sc

ores

for i

dent

ifyin

g eff

ectiv

e te

achi

ng, t

houg

h re

sear

ch u

sing

val

ue-a

dded

mod

els

does

sug

gest

that

te

ache

rs d

iffer

mar

kedl

y in

thei

r con

trib

utio

ns to

stu

-de

nts’

test

sco

re g

ains

. How

ever

, cor

rela

ting

valu

e-ad

d-ed

sco

res

with

teac

her q

ualifi

catio

ns, c

hara

cter

istic

s,

or p

ract

ices

has

yie

lded

uns

tabl

e, in

valid

, and

mix

ed

resu

lts a

nd fe

w s

igni

fican

t find

ings

. Thu

s, it

is o

bvio

us

that

teac

hers

var

y in

effe

ctiv

enes

s, b

ut th

e re

ason

s fo

r th

is a

re n

ot k

now

n.

• A

s al

l gro

wth

mod

els,

VA

Ms

are

pref

er-

able

to “

stat

us”

or a

ttai

nmen

t mod

els

that

m

easu

re

stud

ent p

rofic

ienc

y at

one

poi

nt in

tim

e be

caus

e th

ey in

dica

te c

hang

e in

stu

dent

le

arni

ng o

ver t

ime

• C

an p

rovi

de in

form

atio

n on

som

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

out

com

es

(onl

y th

ose

in te

sted

sub

ject

s an

d gr

ade

leve

ls)

• C

an a

llow

for c

ompa

rison

s be

twee

n sc

hool

s an

d pr

ogra

ms

with

bet

ter a

ccu-

racy

than

with

indi

vidu

al te

ache

rs

• M

ore

likel

y to

mea

sure

impa

ct o

f te

ache

rs v

ersu

s ot

her s

tude

nt a

nd s

choo

l ba

ckgr

ound

fact

ors

than

oth

er k

inds

of

grow

th m

odel

s

• Re

quire

s no

cla

ssro

om v

isits

bec

ause

lin

ked

stud

ent/

teac

her d

ata

can

be a

na-

lyze

d at

a d

ista

nce

• En

tails

litt

le b

urde

n at

the

clas

sroo

m

or s

choo

l lev

el b

ecau

se m

ost d

ata

are

alre

ady

colle

cted

for N

CLB

pur

pose

s

• M

ay b

e us

eful

for i

dent

ifyin

g ou

tsta

nd-

ing

teac

hers

who

se c

lass

room

s ca

n se

rve

as “

lear

ning

labs

” as

wel

l as

stru

gglin

g te

ache

rs in

nee

d of

sup

port

• Es

timat

es o

f tea

chin

g eff

ectiv

enes

s ca

n be

uns

tabl

e fr

om y

ear t

o ye

ar,

whi

ch is

why

mos

t VA

Ms

requ

ire th

ree

or m

ore

year

s of

dat

a

• D

ifficu

lt to

ver

ify th

e ac

cura

cy o

f the

m

easu

res

due

to la

ck o

f tra

nspa

renc

y.

• V

alue

add

ed m

easu

res

are

not a

vail-

able

for t

he m

ajor

ity o

f tea

cher

s

• Re

sults

do

not p

rovi

de s

uffici

ent

info

rmat

ion

on h

ow te

ache

rs c

an im

-pr

ove

thei

r effe

ctiv

enes

s

• C

an le

ad to

fals

e at

trib

utio

ns o

f va

lue—

that

is, i

n te

am- o

r co-

teac

hing

si

tuat

ions

, can

not s

epar

ate

out c

ontr

i-bu

tions

of o

ne te

ache

r or a

noth

er

• Re

quire

s ve

rtic

ally

sca

led

exam

s

• G

row

th m

odel

s ca

n en

cour

age

non-

educ

ativ

e te

st p

rep,

che

atin

g, e

tc.

• G

row

th m

odel

s ca

n di

s-in

cent

iviz

e co

llabo

ratio

n de

pend

ing

on h

ow th

ey

are

used

• M

odel

s ar

e no

t abl

e to

sor

t out

te

ache

r effe

cts

from

cla

ssro

om e

ffect

s.

• V

alue

-add

ed s

core

s ar

e no

t use

ful f

or

form

ativ

e pu

rpos

es b

ecau

se te

ache

rs

lear

n no

thin

g ab

out h

ow th

eir p

ract

ices

co

ntrib

uted

to (o

r im

pede

d) s

tude

nt

lear

ning

• V

alue

-add

ed m

easu

res

are

cont

rove

r-si

al b

ecau

se th

ey m

easu

re o

nly

teac

h-er

s’ c

ontr

ibut

ions

to s

tude

nt a

chie

ve-

men

t gai

ns o

n st

anda

rdiz

ed te

sts

• Re

quire

s no

cla

ssro

om v

isits

bec

ause

lin

ked

stud

ent/

teac

her d

ata

can

be

anal

yzed

at a

dis

tanc

e

Page 41: OEA's Teacher Evaluation & Support System Guidebook

41

Mea

sure

:

Val

ue-A

dded

Mod

els*

Des

crip

tion

:

Val

ue-A

dded

Mod

els

(VA

Ms)

are

cur

-re

ntly

bei

ng u

sed

to d

eter

min

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

s’ te

st

scor

e ga

ins

and

also

are

bei

ng u

sed

as a

rese

arch

tool

(e.g

., de

term

inin

g th

e di

strib

utio

n of

“effe

ctiv

e” te

ache

rs

by s

tude

nt o

r sch

ool c

hara

cter

istic

s).

VAM

’s m

easu

re th

e ga

ins

that

stu

dent

s m

ake

and

adju

st th

ose

gain

s fo

r stu

dent

, te

ache

r, or

sch

ool c

hara

cter

istic

s. T

he

gain

s ar

e in

terp

rete

d as

the

“val

ue”

that

a

teac

her a

dds

beca

use

the

gain

s ar

e pr

esum

ed to

be

net o

f all

othe

r infl

uenc

-es

. Thi

s pr

esum

ptio

n is

muc

h de

bate

d.

Stro

nger

VA

Ms

use

wel

l-de

sign

ed v

erti-

cally

equ

ated

sta

ndar

dize

d ac

hiev

emen

t te

sts

that

mea

sure

rele

vant

con

cept

s an

d te

st s

tude

nts

at le

ast t

hree

tim

es.

Som

e te

sts

are

vert

ical

ly e

quat

ed, w

hich

m

eans

that

a g

iven

sco

re o

n th

e fo

urth

-gr

ade

vers

ion

of a

test

repr

esen

ts th

e sa

me

leve

l of p

erfo

rman

ce a

s th

at s

ame

scor

e on

the

fifth

-gra

de v

ersi

on o

f the

te

st. O

AKS

is n

ot a

ver

tical

ly e

quat

ed

asse

ssm

ent.

Exam

ples

:

• C

LASS

Pro

ject

TIF

Gra

nt re

cipi

ents

, su

ch a

s Sa

lem

-Kei

zer S

choo

l Dis

tric

t, Le

bano

n Sc

hool

Dis

tric

t; al

so M

cMin

-nv

ille

Scho

ol D

istr

ict

• Te

nnes

see

Val

ue-A

dded

Ass

essm

ent

Syst

em (

TVA

AS)

• O

hio’

s Ed

ucat

ion

Val

ue-A

dded

As-

sess

men

t Sys

tem

• D

alla

s V

alue

-Add

ed A

ccou

ntab

ility

Sy

stem

(D

VAA

S)

• Re

sear

cher

s ha

ve w

arne

d ag

ains

t usi

ng v

alue

-add

ed

estim

ates

for h

igh

stak

es p

urpo

ses,

incl

udin

g hi

ring,

fir-

ing,

com

pens

atio

n, a

nd s

umm

ativ

e ev

alua

tions

. (Ba

ker,

et a

l., 2

010

; Bra

un, 2

00

5; B

raun

, Chu

dow

sky

& K

onei

g,

2010

; Cor

cora

n, 2

020

; Kup

erm

intz

, 20

03)

• La

ck o

f con

sens

us in

the

rese

arch

com

mun

ity a

bout

w

heth

er V

AM

s ca

n ac

cura

tely

isol

ate

the

effec

ts o

f a

sing

le te

ache

r, es

peci

ally

ove

r an

exte

nded

per

iod.

(G

olds

chm

idt,

et a

l., 2

00

5; K

oede

l & B

etts

, 20

09;

Kan

e &

Sta

iger

, 20

08)

• Te

ache

r inp

ut m

ay b

e an

impo

rtan

t infl

uenc

e on

ac

hiev

emen

t, bu

t it i

s no

t the

onl

y in

fluen

ce.

(Bra

un,

200

5)

• VA

M n

umbe

rs d

o no

t offe

r for

mat

ive

feed

back

abo

ut

way

s to

impr

ove

inst

ruct

ion

or c

lass

room

effe

ctiv

enes

s.

(Mill

er, 2

00

9; R

owan

, Cor

rent

i, &

Mill

er, 2

00

2)

• VA

Ms

fail

to a

ccur

atel

y re

flect

the

limita

tions

of p

ar-

ticul

ar te

sts

both

for m

easu

ring

the

full

rang

e of

des

ired

know

ledg

e an

d sk

ills

and

for m

easu

ring

lear

ning

gai

ns.

(New

ton,

Dar

ling-

Ham

mon

d, H

aert

el &

Tho

mas

, 20

10)

• St

uden

ts a

re n

ot n

eces

saril

y as

sign

ed to

cla

ssro

oms

and

teac

hers

on

a ra

ndom

bas

is. T

his

has

maj

or ra

mi-

ficat

ions

on

man

y VA

Ms.

Thi

s w

ill s

ever

ely

impa

ct th

e va

lidity

and

relia

bilit

y of

a v

alue

-add

ed s

core

. Acc

urat

e VA

Ms

requ

ire c

ompa

rabi

lity

acro

ss d

emog

raph

ics

of a

n in

divi

dual

teac

her’s

cla

ssro

om o

r sch

ool s

ettin

g. (

Har

-ris

, 20

09)

• VA

Ms

are

expe

nsiv

e to

ado

pt a

nd m

aint

ain.

Mos

t sc

hool

dis

tric

ts c

anno

t affo

rd to

upk

eep

the

rigor

ous

data

requ

irem

ents

or t

he d

egre

e of

hum

an re

sour

ces

and

psyc

hom

etric

exp

ertis

e re

quire

d. (G

olds

chm

idt,

200

5; H

arris

, 20

09)

• Li

ttle

is k

now

n ab

out t

he v

alid

ity o

f val

ue-a

dded

sc

ores

for i

dent

ifyin

g eff

ectiv

e te

achi

ng, t

houg

h re

sear

ch u

sing

val

ue-a

dded

mod

els

does

sug

gest

that

te

ache

rs d

iffer

mar

kedl

y in

thei

r con

trib

utio

ns to

stu

-de

nts’

test

sco

re g

ains

. How

ever

, cor

rela

ting

valu

e-ad

d-ed

sco

res

with

teac

her q

ualifi

catio

ns, c

hara

cter

istic

s,

or p

ract

ices

has

yie

lded

uns

tabl

e, in

valid

, and

mix

ed

resu

lts a

nd fe

w s

igni

fican

t find

ings

. Thu

s, it

is o

bvio

us

that

teac

hers

var

y in

effe

ctiv

enes

s, b

ut th

e re

ason

s fo

r th

is a

re n

ot k

now

n.

• A

s al

l gro

wth

mod

els,

VA

Ms

are

pref

er-

able

to “

stat

us”

or a

ttai

nmen

t mod

els

that

m

easu

re

stud

ent p

rofic

ienc

y at

one

poi

nt in

tim

e be

caus

e th

ey in

dica

te c

hang

e in

stu

dent

le

arni

ng o

ver t

ime

• C

an p

rovi

de in

form

atio

n on

som

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

out

com

es

(onl

y th

ose

in te

sted

sub

ject

s an

d gr

ade

leve

ls)

• C

an a

llow

for c

ompa

rison

s be

twee

n sc

hool

s an

d pr

ogra

ms

with

bet

ter a

ccu-

racy

than

with

indi

vidu

al te

ache

rs

• M

ore

likel

y to

mea

sure

impa

ct o

f te

ache

rs v

ersu

s ot

her s

tude

nt a

nd s

choo

l ba

ckgr

ound

fact

ors

than

oth

er k

inds

of

grow

th m

odel

s

• Re

quire

s no

cla

ssro

om v

isits

bec

ause

lin

ked

stud

ent/

teac

her d

ata

can

be a

na-

lyze

d at

a d

ista

nce

• En

tails

litt

le b

urde

n at

the

clas

sroo

m

or s

choo

l lev

el b

ecau

se m

ost d

ata

are

alre

ady

colle

cted

for N

CLB

pur

pose

s

• M

ay b

e us

eful

for i

dent

ifyin

g ou

tsta

nd-

ing

teac

hers

who

se c

lass

room

s ca

n se

rve

as “

lear

ning

labs

” as

wel

l as

stru

gglin

g te

ache

rs in

nee

d of

sup

port

• Es

timat

es o

f tea

chin

g eff

ectiv

enes

s ca

n be

uns

tabl

e fr

om y

ear t

o ye

ar,

whi

ch is

why

mos

t VA

Ms

requ

ire th

ree

or m

ore

year

s of

dat

a

• D

ifficu

lt to

ver

ify th

e ac

cura

cy o

f the

m

easu

res

due

to la

ck o

f tra

nspa

renc

y.

• V

alue

add

ed m

easu

res

are

not a

vail-

able

for t

he m

ajor

ity o

f tea

cher

s

• Re

sults

do

not p

rovi

de s

uffici

ent

info

rmat

ion

on h

ow te

ache

rs c

an im

-pr

ove

thei

r effe

ctiv

enes

s

• C

an le

ad to

fals

e at

trib

utio

ns o

f va

lue—

that

is, i

n te

am- o

r co-

teac

hing

si

tuat

ions

, can

not s

epar

ate

out c

ontr

i-bu

tions

of o

ne te

ache

r or a

noth

er

• Re

quire

s ve

rtic

ally

sca

led

exam

s

• G

row

th m

odel

s ca

n en

cour

age

non-

educ

ativ

e te

st p

rep,

che

atin

g, e

tc.

• G

row

th m

odel

s ca

n di

s-in

cent

iviz

e co

llabo

ratio

n de

pend

ing

on h

ow th

ey

are

used

• M

odel

s ar

e no

t abl

e to

sor

t out

te

ache

r effe

cts

from

cla

ssro

om e

ffect

s.

• V

alue

-add

ed s

core

s ar

e no

t use

ful f

or

form

ativ

e pu

rpos

es b

ecau

se te

ache

rs

lear

n no

thin

g ab

out h

ow th

eir p

ract

ices

co

ntrib

uted

to (o

r im

pede

d) s

tude

nt

lear

ning

• V

alue

-add

ed m

easu

res

are

cont

rove

r-si

al b

ecau

se th

ey m

easu

re o

nly

teac

h-er

s’ c

ontr

ibut

ions

to s

tude

nt a

chie

ve-

men

t gai

ns o

n st

anda

rdiz

ed te

sts

• Re

quire

s no

cla

ssro

om v

isits

bec

ause

lin

ked

stud

ent/

teac

her d

ata

can

be

anal

yzed

at a

dis

tanc

e

Mea

sure

:

Nor

mat

ive

Gro

wth

Mod

els*

Des

crip

tion

:

Nor

mat

ive

Gro

wth

Mod

els

com

pare

gr

owth

in s

tude

nt a

chie

vem

ent t

o th

e am

ount

of g

row

th m

ade

by a

repr

e-se

ntat

ive

popu

latio

n of

stu

dent

s on

th

e sa

me

test

. A v

ertic

al s

cale

is n

ot

nece

ssar

y.

Exam

ples

:

• M

assa

chus

etts

Stu

dent

Gro

wth

Pe

rcen

tile

(SG

P) (

not c

urre

ntly

use

d to

mea

sure

teac

hing

effe

ctiv

enes

s, b

ut

rath

er to

indi

cate

wha

t pro

fess

iona

l gr

owth

trac

k a

teac

her w

ill b

e on

)

• N

ew Je

rsey

Stu

dent

Gro

wth

Per

cent

ile

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

” an

d re

sear

ch in

VA

M s

ectio

n•

Are

pre

fera

ble

to “

stat

us”

mod

els

beca

use

they

indi

cate

cha

nge

in s

tude

nt

lear

ning

ove

r tim

e

• C

an p

rovi

de in

form

atio

n on

som

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

out

com

es

• C

an a

llow

for c

ompa

rison

s be

twee

n te

ache

rs

• U

nlik

e ot

her g

row

th m

odel

s, th

ese

do

not r

ely

on v

ertic

ally

-sca

led

exam

s

• C

anno

t be

com

pute

d fo

r tea

cher

s in

un

test

ed g

rade

s an

d su

bjec

ts

• C

an le

ad to

fals

e at

trib

utio

ns o

f va

lue—

that

is, i

n te

am- o

r co-

teac

hing

si

tuat

ions

, can

not s

epar

ate

out c

ontr

i-bu

tions

of o

ne te

ache

r or a

noth

er

• G

row

th m

odel

s ca

n en

cour

age

non-

educ

ativ

e te

st p

rep,

che

atin

g, e

tc.

• C

anno

t ade

quat

ely

cont

rol f

or s

tu-

dent

bac

kgro

und

char

acte

ristic

s

• G

row

th m

odel

s ca

n di

s-in

cent

iviz

e co

llabo

ratio

n de

pend

ing

on h

ow th

ey

are

used

Mea

sure

:

Cat

egor

ical

Gro

wth

Mod

els*

Des

crip

tion

:

Cat

egor

ical

Gro

wth

Mod

els

calc

ulat

e st

uden

t gro

wth

bas

ed o

n ch

ange

s in

pe

rfor

man

ce c

ateg

ory

plac

emen

t (e.

g.,

from

“be

ginn

ing”

to “

profi

cien

t”)

from

ye

ar to

yea

r. C

hang

es in

all

poss

ible

ca

tego

ry p

lace

men

ts a

re ju

dged

sub

jec-

tivel

y an

d ea

ch is

ass

igne

d a

valu

e th

at

inde

xes

its im

port

ance

.

Exam

ples

:

• Fl

orid

a V

alue

Tab

les

• M

inne

sota

Gro

wth

Mod

el

Rese

arch

fort

hcom

ing

– se

e “E

xam

ples

”•

Are

pre

fera

ble

to “

stat

us”

mod

els

beca

use

they

indi

cate

cha

nge

in s

tude

nt

lear

ning

ove

r tim

e

• C

an p

rovi

de in

form

atio

n on

som

e te

ach-

ers’

con

trib

utio

ns to

stu

dent

out

com

es

• C

an a

llow

for c

ompa

rison

s be

twee

n te

ache

rs

• A

re m

ore

easi

ly u

nder

stan

dabl

e by

var

i-ou

s st

akeh

olde

rs th

an V

AM

s

• C

anno

t ade

quat

ely

cont

rol f

or s

tu-

dent

’s b

ackg

roun

d ch

arac

teris

tics

• G

row

th m

odel

s m

ay e

ncou

rage

non

-ed

ucat

ive

test

pre

p, c

heat

ing,

etc

.

• G

row

th m

odel

s ca

n di

s-in

cent

iviz

e co

llabo

ratio

n de

pend

ing

on h

ow th

ey

are

used

* N

ote

that

one

opt

ion

for u

sing

mea

sure

s of

stu

dent

gro

wth

in te

ache

r eva

luat

ion

is to

cal

cula

te s

choo

l-wid

e st

uden

t gro

wth

usi

ng a

ny o

f the

sta

tistic

al m

odel

s de

scrib

ed in

this

tabl

e an

d us

e th

at a

s on

e of

mul

tiple

mea

sure

s of

teac

hers

’ con

trib

utio

ns to

sch

ool a

nd s

tude

nt o

utco

mes

. Dep

endi

ng o

n ho

w s

choo

l-wid

e st

uden

t gro

wth

dat

a is

use

d (e

.g. i

nfor

m te

achi

ng p

ract

ice)

, it c

ould

hav

e th

e po

-te

ntia

l ben

efit o

f mot

ivat

ing

all t

each

ers

in th

e sc

hool

to w

ork

with

thei

r stu

dent

s on

par

ticul

ar k

now

ledg

e an

d sk

ills

(for

exa

mpl

e, a

ll te

ache

rs c

an p

rom

ote

stud

ent w

ritin

g in

thei

r cla

sses

). H

owev

er

if us

ed in

hig

h-st

ake

situ

atio

ns (e

.g.,

cont

inua

tion

of e

mpl

oym

ent)

, it m

ay p

ut u

ndue

pre

ssur

e on

teac

hers

who

se p

rimar

y re

spon

sibi

lity

is th

e te

sted

sub

ject

are

a.

Page 42: OEA's Teacher Evaluation & Support System Guidebook

42

MEA

SURE

S OF

STU

DENT

LEAR

NING

Mea

sure

/ D

escr

iptio

n /

Exam

ples

Rese

arch

Stre

ngth

sCa

utio

nsD

istr

ict-

dete

rmin

ed m

easu

res o

f stu

dent

lear

ning

, com

para

ble

acro

ss g

rade

or s

ubje

ct d

istr

ict-

wid

e (i.

e., c

omm

on fo

rmat

ive

and

sum

mat

ive

asse

ssm

ents

)

Mea

sure

:

Com

mon

For

mat

ive

Ass

essm

ents

Des

crip

tion:

Ass

essm

ents

cre

ated

by

scho

ol-d

istr

icts

(or p

oten

tially

at

the

build

ing

leve

l) th

at a

re c

ompa

rabl

e ac

ross

sub

ject

s an

d gr

ades

.

Exam

ples

:

Acc

ordi

ng to

the

Mas

sach

uset

ts’ R

TTT

appl

icat

ion,

MA

will

w

ork

to d

evel

op d

istr

ict-

base

d as

sess

men

ts th

at a

re c

ompa

-ra

ble

acro

ss s

ubje

cts

and

grad

es.

Rese

arch

fort

hcom

ing

– se

e “E

x-am

ples

Dou

g Re

eves

“L

earn

ing

By D

oing

” - s

ugge

sts

not u

sing

form

ativ

e as

sess

men

ts

for t

each

er e

valu

atio

n or

gra

ding

st

uden

ts

• In

ter-

rate

r rel

iabi

lity

•Mul

tiple

opt

ions

• Ea

sy to

com

pare

resu

lts a

mon

g te

ache

rs

• W

orks

wel

l with

PLC

’s

• M

ore

equi

tabl

e fo

r stu

dent

s

•May

sav

e te

ache

rs’ t

ime

by c

reat

ing

effici

enci

es

• C

an b

e m

ore

auth

entic

ass

essm

ents

of s

tude

nt

lear

ning

than

sta

ndar

dize

d te

sts

and

ther

efor

e m

ore

valid

ass

essm

ents

of t

each

ers’

con

trib

utio

ns

to s

tude

nt le

arni

ng

• C

an p

rovi

de im

port

ant f

orm

ativ

e an

d su

mm

ativ

e in

form

atio

n ab

out a

teac

her’s

pra

ctic

e

•Can

be

time

cons

umin

g to

cre

ate

• M

ust p

lan

ahea

d

• N

eed

cons

ensu

s w

ith o

ther

teac

hers

te

achi

ng c

lass

, if t

here

are

any

, in

crea

tion

of a

sses

smen

t

• G

oal i

s st

uden

t ach

ieve

men

t, no

t in

tend

ed fo

r tea

cher

eva

luat

ion

• Va

lidity

and

relia

bilit

y no

t wel

l est

ab-

lishe

d

• Ve

ry d

ifficu

lt to

sta

ndar

dize

and

ther

e-fo

re d

ifficu

lt to

est

ablis

h co

mpa

rabi

lity

and

relia

bilit

y

• Ps

ycho

met

ric p

rope

rtie

s of

suc

h as

-se

ssm

ents

are

not

wel

l und

erst

ood ADDITIONAL CHALLENGES

WITH USING STUDENT LEARNING AND GROWTH DATA

It is also important to consider the following chal-lenges as you plan to use any measures of student learning and growth.

Teacher preparation and ongoing professional de-velopment for measuring student learning growth is limited

Most principals, support providers, instructional managers, and coaches are not adequately prepared to make judgments about teachers’ contributions to student learning and growth

They need to know how to: Evaluate the appropriateness of various mea-

sures of student learning for use in teacher evaluation

Work closely with teachers to select appro-priate student growth measures and ensure that they are using them correctly and con-sistently.

The use of student learning goals based

on multiple measures of student learning is

a more sound solution, rather than directly

connecting individual teachers to their

students’ statewide assessment (OAKS)

scores.

Page 43: OEA's Teacher Evaluation & Support System Guidebook

43

CONNECTING STATEWIDE TEST SCORES TO TEACHER EVALUATION: AN EDUCATIONALLY SOUND APPROACH?

The primary goal for a teacher is to facilitate student learning. Toward that end, most teachers routinely

utilize student learning data, includ-ing achievement data, to improve their instruction and practice. So it might seem to follow that basing teacher evaluations on student test scores would make sense. Not so fast. In reality, when it comes to measuring teacher performance, a battery of research has demonstrated that test scores are not very reliable at all. As such, test scores should never be used to make high-stakes personnel decisions, nor should student tests, such as statewide, summative, standardized assessments, be utilized as the sole measure of teaching ef-fectiveness in a summative evaluation or the sole measure of student learning.

If you are integrating student assess-ments as one measure in your evaluation system, remember there are the good, the bad and the ugly. What are some things to watch out for when it comes to utiliz-ing measures based on student learning and growth? First, be mindful of whether achievement data is derived from de-velopmentally appropriate tests. If it is not developmentally appropriate, it is probably not a valid measure. Second, seek reliability. If the measure of student learning is unreliable, it is not valid for evaluating teaching effectiveness. Last, steer clear of student assessments where data is used only in a summative fashion for teacher evaluation and not for purpos-es of improvement and growth of teaching practice.

Just as it is vital to utilize multiple mea-sures of teaching effectiveness to look holistically at of the art and practice of teaching, when using student data one should never rely solely on test scores. It is imperative to use multiple measures of student learning including other forms of evidence of learning like student portfo-lios and work samples.

THE ULTIMATE GOAL IS TO IMPROVE TEACHER PRACTICE AND STUDENT LEARNING AND GROWTH

The primary goal of a quality Teacher Evaluation and Support System (TESS) is to promote professional growth and develop teacher capacity and to provide tools to improve the skills, knowledge and craft required of good teaching. Personnel decisions are only a very small fraction of the entire aim of said systems. Thus, if there are measures within a TESS that do not have a valid research foundation showing they work to improve teaching or learning, they should not be included in a system.

Additionally, there are many ways to incorporate student learning data into a Teacher Evaluation and Support System that will not inadvertently punish teach-ers for those flaws, but help teachers and educators to strengthen their practice and performance and to improve student learning. A sound approach to incorporat-ing student data is through the use of stu-dent learning goals. Rather than directly connecting individual teachers to their students’ statewide assessment scores, setting student learning growth goals provides an opportunity to use multiple measures of student learning to provide a more holistic picture of teacher practice and performance.

USING STATEWIDE TEST SCORES: WHY IT REMAINS AN EDUCATIONALLY UNSOUND APPROACH

• Reliable and Valid Data: Oregon’s statewide assessment system (OAKS) has strand level data that is often so unreliable, it should not be used as the sole measure for even large-scale, building-level program decisions. Many school districts use additional forma-tive assessments for the purpose of making programmatic changes. Even so, if statewide test data is not reliable or valid for program-level decisions, it should not be used for individual teacher evaluation.

• Test Scores Show Patterns Over Time: While three to four more years of test scores might reliably show student growth or lack thereof, a single year of scores is not a reliable measure of teaching effectiveness.

• Calibration for Growth: In order for three to four years of test scores to be viewed, the assessment itself needs to be vertically scaled so educators can compare one year to the next. Oregon’s assessment is not currently calibrated for this purpose. In other words, the assessment is not designed to show student growth from year to year. Until our state assessment is appropriately scaled, using state test scores for this purpose is invalid.

• Context Matters: Many factors outside of school influence student achievement and therefore affect stu-dent test scores. These include home support, school attendance, family in-come level, and parents’ level of educa-tion. Moreover, there are other factors inside a school, which a teacher does not control, that can greatly impact a student’s academic and social readiness to pass a test. These include principals, peer groups, curricula, the school build-ing/environment, supplies available, courses offered, etc. While the teacher certainly plays the largest in-school role for predicting student success, there are far too many other variables in play to directly link “student outcomes” solely to their “teacher inputs” (see CD-ROM, Research, Teacher Effectiveness on Student Learning).

• Multiple Valid Points of Data: Test scores do not speak for themselves. Teacher Evaluation and Support Systems that incorporate student data need multiple valid points of data from both a quantitative and qualitative perspective to truly reflect the whole child and the process of learning. OAKS scores do not accurately reflect many things that teachers do to produce qual-itative and other outcomes that cannot be captured on a multiple choice test

Using Measures of Student LearningVI.

as a Measure of Teaching Effectiveness

Page 44: OEA's Teacher Evaluation & Support System Guidebook

44

(such as higher-order thinking skills). USING STATEWIDE TEST SCORES: IT IS NOT A LOGISTICALLY POSSIBLE SOLUTION

• Limited Tested Grades: The state level assessment is given in grades 3-8 and 11. It is not equitable that some teachers may be evaluated on the less reliable statewide standardized assess-ment, while others may not.

• Limited Tested Subjects: The state assessment addresses selected subject areas and does not cover the majority of the content in the school curriculum. Reading, mathematics, science, and writ-ing are the only subject areas assessed on the statewide tests, and they are not con-sistently assessed in every grade. Again, this creates a systemic inequity.

• Over 60% of Teaching Staff are Not Connected to Statewide Assess-ments: Combining the grades where the state assessments are not tested and the non-tested subject areas, over 60% of certified teachers/specialists are not even connected with statewide test scores making a comprehensive system using state test scores for teacher evalu-ations logistically impossible.

POSSIBLE RAMIFICATIONS OF USING STATEWIDE TEST SCORE DATA

• Less Collaboration and More Com-petition: If teachers are being evalu-ated based on state test scores, they are less likely to collaborate and help their colleagues. More and more research suggests that teacher collaboration is one of the best forms of professional development, and it is also indicative of student learning gains. Why would we want an educational system that does not fully promote collaboration?

• Teaching to the Assessment: If student test scores on a single assess-ment becomes the basis for teacher evaluation, then the test will become the major instructional focus, crowding out broader learning to an even greater extent.

• The Campbell Effect: Generally speak-ing, the Campbell Effect states that when test scores become the goal of the

teaching process, they lose their value as indicators of educational status and dis-tort the educational process in undesir-able ways. In other words, the pressure to have students score well on a single test for teacher evaluation becomes so intense that it potentially leads to un-scrupulous practices including:

★ Cheating on the test by both stu-dents and teachers

★ Data manipulation ★ Distorts education by narrowing

the curriculum ★ Distorts education by teaching to

the testThe Campbell Effect has been demon-strated in public and private sectors, and its most detrimental effect is demoral-izing the workforce charged with carrying out the assessments.

DESIGNING A TEACHER EVALUATION AND SUPPORT SYSTEM THAT SUPPORTS GROWTH

In designing a way to measure teaching effectiveness, it is important to keep a couple of things in mind. First, the goal of a new Teacher Evaluation and Support System is improvement. The main goal of evaluation is pointing out areas where teachers need improvement, then provid-ing support so they can grow and develop. Creating a system that promotes growth is an educationally sound solution to increasing teacher quality across Oregon that is supported by research and con-nected to good practice. Second, a well-rounded Teacher Evaluation and Support System includes multiple valid points of data to determine teaching effectiveness, collected via multiple valid measures linked to improving student performance including multiple valid measures of student learning. Multiple valid measures provide opportunities for triangulation of data and provide a much stronger and holistic representation of teaching ef-fectiveness. Other considerations in an effective Teacher Evaluation and Support System include: • Collaboration as a key principle in

designing and implementing an evalua-tion system

• A clear and consistent administrator training and professional development component, including calibration and inter-rater reliability checks

• A comprehensive beginning teacher/ induction program

• A comprehensive mentoring system including mentor release time for col-laboration with mentees

• Ongoing, focused and job-embedded professional development to aid improving instructional practice and student learning, following the Learn-ing Forward standards (see CD-ROM)

• Teachers demonstrating use of student achievement to inform and improve instruction

• Teachers and administrators being trained in demonstrating the impact they have had on student learning using multiple pieces of evidence

• Transparent and timely access to data • Multiple opportunities for classroom

observations by highly-trained admin-istrators

• Self-assessment and self-reflection em-bedded in teacher evaluation systems

• Collaboration time with colleagues and principal focused on instructional practice and student learning

• Release time for employees to observe mentor(s) and other accomplished col-leagues

• Clear link to the InTASC Standards in the evaluation criteria and application of the standards across the TESS

• A differentiated evaluation rubric that measures growth of teaching practices against teaching standards

• Additional administrator support to help implement a growth oriented evaluation system.

When using student data, it is important to not solely use statewide test scores for formative or summative teacher evalua-tion. It is imperative to use multiple mea-sures of student learning to gain a more holistic assessment of what students know, just as it is vital to utilize multiple mea-sures of teaching effectiveness to gauge the entirety of the art and science of teaching.

Adapted from Washington Education Association

Using Measures of Student LearningVI.

as a Measure of Teaching Effectiveness

Page 45: OEA's Teacher Evaluation & Support System Guidebook

45

I. Value-Added Models Used to Inform Instruction and Programs at the School Level

Use of VAM Value of UseVAM as one measure with other multiple valid measures of student learning to be used in a whole school Formative Evaluation process, designed to inform and improve practice

Low-Stakes, Moderate Efficacy• Reliability & validity in increased because other mul-

tiple valid measures are in use (Goe, 2010)

VAM as sole measure of student learning to be used as a whole school Forma-tive Evaluation process, designed to inform and improve practice

Low Stakes, Low Efficacy• VAM is a limited diagnostic tool, in part because in

cannot demonstrate in-school variables on perfor-mance

VAM to be used as a “trigger” to examine a school’s performance more closely, but not to be used as part of a final or Summative Evaluation, or used for high-stakes decision making

Moderate Stakes, Low Efficacy

II. Value-Added Models Used to Inform Instruction at the Individual Teacher Level

Use of VAM Value of UseVAM as one measure with other multiple valid measures of student learning to be used in an individual teacher Formative Evaluation process, designed to inform and improve practice

Low-Stakes, Low Efficacy• Although reliability & validity is increased because

other multiple valid measures are in use (Goe, 2010), VAM is not designed to give specific formative feed-back at the individual level (essentially, its use in this regard is moot – all useful feedback would be coming from the other multiple valid measures).

VAM as sole measure of student learning to be used in an individual teacher Formative Evaluation process, designed to inform and improve practice

Moderate Stakes, Low Efficacy

VAM to be used as a “trigger” to examine a teacher’s performance more closely, but not to be used as part of a final or Summative Evaluation, or used for high-stakes decision making such as hiring, firing, or compensation, etc.

Moderate -High Stakes, Low Efficacy

III. Value-Added Models Used to Evaluate Individual PerformanceCareful consideration should be given to the “weight” of VAM against other measures.

Use of VAM Value of UseVAM to be used as one measure with other multiple valid measures of stu-dent learning and student learning then becoming one of multiple measures of teaching effectiveness – a measure within multiple measures within multiple measures.

CAUTION: High Stakes, Low Efficacy

VAM to be used as a significant percentage of a Summative Evaluation or other high-stakes decision.

EXTREME CAUTION: High Stakes, Low Efficacy

VAM to be used as a sole measure for a Summative Evaluation system, or sole measure for high-stakes decisions such as hiring, firing, compensation, etc.

STOP: High Stakes, Low Efficacy

OEA’S ANALYSIS OF USAGE OF A VALUE-ADDED MODEL FOR DIFFERENT PURPOSES

Value Added Models VII.

Page 46: OEA's Teacher Evaluation & Support System Guidebook

46

A value-added model is a stu-dent growth model utilizing achievement data (namely,

OAKS scores or other state-wide summative assessments) and ad-ditional student background data to attempt to isolate the specific effects of the teacher, school, or program on student academic achievement progress. This data is then used to determine whether or not a student’s average change in performance is meeting a pre-de-termined growth target. In many Oregon school districts utilizing value-added models (VAMs), a school-level model is being uti-lized, which attributes student outcomes to the school as a whole, and not to individual teachers.

Proponents of value-added models as-sert that objective data about whether students have learned must be taken into consideration. They contend that, despite its flaws, a VAM is the best model available to show whether a school’s test scores are improving over time. There are questions, however, about the reliability and validity of these models and the esti-mates they produce. Therefore, caution should be taken when using a VAM as a basis for high-stakes decisions related to individual teacher performance or pay.

Ultimately, if the purpose of an evaluation system is to improve teaching and learn-ing, then it is important that all measures and models provide diagnostic insight that will help improve the teaching and learning processes. However, VAMs are not diagnostic tools designed to facilitate the improvement of teaching and learn-ing, and thus conflict with this goal.

VALUE-ADDED MODELS: WHAT TO CONSIDER • Researchers have warned against using

value-added estimates for high stakes purposes, including hiring, firing, com-pensation, and summative evaluations. (Baker, et al., 2010; Braun, 2005; Braun, Chudowsky & Koneig, 2010; Corcoran, 2020; Kupermintz, 2003)

• There is a lack of consensus in the research community about whether VAMs can accurately isolate the effects of a single teacher, especially over an extended period and distinguished from other classroom and in-school effects. (Goldschmidt, et al., 2005; Koedel & Betts, 2009; Kane & Staiger, 2008)

• Teacher input may be an important influence on achievement, but it is not the only influence. (Braun, 2005)

• VAM numbers don’t say much about achievement in that they do not tell us why some classrooms are more effective, why some schools are more effective, or why some teachers are more effective. They do not offer forma-tive feedback about ways to improve instruction. (Miller, 2009; Rowan, Cor-renti, & Miller, 2002)

• VAMs fail to accurately reflect the limitations of particular tests both for measuring the full range of desired knowledge and skills and for measuring learning gains. (Newton, Darling-Ham-mond, Haertel & Thomas, 2010)

• Students are not necessarily assigned to classrooms and teachers on a random basis. This has major ramifications on many VAMs. This will severely impact the validity and reliability of a value-added score. Accurate VAMs require comparability across demographics of an individual teacher’s classroom or school setting. (Harris, 2009)

• VAMs are expensive to adopt and maintain. Without continuing outside sources of funding, most school dis-tricts cannot afford to upkeep the rigor-

ous data requirements or the degree of human resources and psychometric expertise required. (Goldschmidt, 2005; Harris, 2009)

• The potential of having teachers engage in a criteria-referenced measure (like VAM) which measures and ranks teach-ers against one another clashes with the priority of having a norm-referenced evaluation system (based on InTASC Standards).

• NEA’s Policy on Teacher Evaluations: Teacher evaluations must be com-prehensive – based on multiple valid indicators to provide teachers with clear and actionable feedback to en-hance their practice – and must include all three of the following components: 1) Indicators of Teacher Practice, 2) Indicators of Teacher Contribution and Growth, and 3) Indicators of Contribu-tion to Student Learning and Growth. Such indicators must be authentic, reflect that there are multiple factors that impact a student’s learning beyond a teacher’s control, and may include the “various indicators” chosen by local or state affiliates. Unless such tests are shown to be developmentally appropri-ate, scientifically valid and reliable for the purpose of measuring both student learning and a teacher’s performance, such tests may not be used to sup-port any employment action against a teacher and may be used only to provide non-evaluative formative feedback.

• OEA believes: The strongest indica-tors of contribution to student learning and growth should come from multiple sources. And while using tests that are shown to be developmentally appropri-ate, scientifically valid and reliable for the purpose of measuring both student learning and a teacher’s performance is a better alternative, at this time no such instrument exists!

Value Added ModelsVII.

Is there Added Value in Value-Added Models?

Page 47: OEA's Teacher Evaluation & Support System Guidebook

47

Value Added ModelsVII.

1. School boards, administrators and unions/associations should review vari-ous models of incentive compensation plans, including research about their effectiveness, before developing a plan at the local level.

2. School boards, administrators and unions/associations should work together to build ongoing community and stakeholder support for both the incentive compensation plan as well as the necessary funding.

3. School boards, administrators and unions/associations should work together to develop and implement the plan utilizing collective bargain-ing where it exists. In locations where collective bargaining does not exist, teachers who would be using the new system should indicate their support for the program.

4. In the implementation of the incentive compensation plan, teachers should be provided assistance, including time, curriculum and professional develop-ment to increase student achievement.

5. The foundation of incentive com-pensation plans shall be profession-al-level base salaries.

6. Funding for the plan shall be ad-equate and sustainable.

7. The plan and its requirements should be transparent, easily under-stood and uniformly implemented.

8. A detailed implementation plan, with agreed-upon benchmarks and timelines, should be developed.

9. The incentive compensation plan should be based on a multi-factor ap-proach (e.g., teacher evaluations, stu-dent performance growth, specific goals set by the teachers and man-agement, increased responsibilities, assessments of student learning) that is research-based and improves student achievement.

10. All employees who meet the crite-ria for the incentive compensation plan should be compensated accord-ingly, and incentive compensation plans should foster collaboration not competition.

11. Evaluations, if a factor in incentive compensation plans, should be fair, of high quality and rigorous, and shall take into account multiple valid measures of student progress.

Please see OEA’s Professional Pay Tool-kit on the CD-ROM.

Shared Principles for Use of Value-Added Models Within Alternative Compensation Models

If your district and association leadership are engaged in a CLASS Project or have a Federal Teacher Incentive Fund grant that requires

a linkage between evaluation and compensation, below are some guiding principles that can help frame the conversation for developing and imple-menting an incentive compensation plan. These principles were developed by the National Educa-tion Association (NEA), in conjunction with the American Association of School Administrators (AASA) and the National School Boards Associa-tion (NSBA):

Page 48: OEA's Teacher Evaluation & Support System Guidebook

4848

Page 49: OEA's Teacher Evaluation & Support System Guidebook

49

STEP 1: SELF-REFLECTIONBased on the standards of professional practice, the first step of an evaluation system is self-reflection. The educa-tor reflects on and assesses his/her professional practice and analyzes the learning and growth of his/her students in preparation for goal setting.

STEP 2: GOAL SETTING (Student growth goals and professional goals)Based on the self-assessment, the educator identifies goals aligned with the standards of professional practice that en-compass both practice and impact on student learning. The educator sets both professional practice goals and student learning goals. SMART goals and/or learning targets are used as a tool for effective goal-setting.

STEP 3: OBSERVATION AND COLLECTION OF EVIDENCE (Multiple measures)The educator and evaluator collect evidence using multiple measures regarding professional practice, professional responsibilities, and student learning to inform progress throughout the process of evaluation.

STEP 4: FORMATIVE ASSESSMENT/EVALUATION (Analysis of evidence, professional conversations and profes-sional growth)The evaluator and educator review the educator’s progress toward goals and/or performance against standards. This step includes three interdependent and critical parts: analy-sis of evidence, professional conversations, and professional growth. Both the educator and the observer analyze the evidence leading into a collaborative professional conversa-tion. Feedback through professional conversations pro-motes awareness of growth that has occurred, and highlights professional growth needs. These conversations help the educator make adjustments in his/her practice and select relevant professional learning opportunities.

STEP 5: SUMMATIVE EVALUATIONThis step is the culmination of multiple formative observa-tions, reflections, professional conversations, etc. Evaluator assesses the educator’s performance against the standards of professional practice, attainment of student learning goals, and attainment of professional practice goals.

Steps in an Evaluation and Professional Growth Cycle

Self Reflection

Goal Setting

Observation/

Collection of Evidence

Formative Assessment/Evaluation

Summative

Evaluation

from ODE Framework for Teacher Evaluation & Support Systems

Evaluation ProceduresVIII.

Page 50: OEA's Teacher Evaluation & Support System Guidebook

50

Peer Assistance programs can and should be part of a larger system of support and growth for all teach-

ers in the district. The formative nature of peer assistance is a critical and often neglected piece of the full continuum of professional learning as defined by the Learning Forward Standards for Profes-sional Learning (www.learningforward.org ). Those standards are based on the profession’s best thinking and research about how to promote outcomes so that every educator engages in effective professional learning every day so every student achieves. Learning Forward Standards for Professional Learning in-cludes developing systems of learning and growth that speak to the following critical elements:

• Learning communities• Leadership• Resources• Data• Learning Designs• Implementation• Outcomes

Peer Assistance programs can offer struc-tured and nuanced formative feedback to teachers about their performance as well as create a community of learners that in-form the professional development that is offered both individually and collectively in a school or district. A PA program must align to the larger evaluation and support system being developed by the district and the association in order to be effective and sustainable. The greatest potential benefit is that instructional support aligns with the evaluation system and is deliv-ered by trained peers whose interest is improvement.

It is important to clearly articulate the roles and responsibilities of peer coaches so that the confidentiality firewall is protected. Otherwise, you not only risk jeopardizing the critical trust between the peer coach and their charge, but precari-ously venture into summative evaluation territory. In the state of Oregon, existing statutes and administrative rules separate a peer coach from a summative evalua-tor. (See PA (Peer Assistance) Versus PAR (Peer Assistance Review))

Peer assistants can be called many things: mentors,

lead teachers,critical friends,

peer coaches, instructional coaches

Their principal role is to guide and mentor teachers through constructive feedback tied directly to the standards being used in the evaluation system. Their primary goal of that feedback is to support the growth and development of their peers (thereby improving retention) but ultimately it can serve to elevate the professional learning of every teacher in the system. PAs can assist not only probationary teachers but also career teachers who may need assistance or are developing new skills. Besides observing teachers and providing feedback specific to the standards in the evaluation system, they might also be assisting their peers to do the following:

• Identify teaching goals• Outline and seek professional growth

activities• Provide support and induction with

lesson plans, parent communication,

assessing student learning, differen-tiating instruction, finding necessary resources, etc.

Peer assistants should be carefully chosen and purposefully trained to perform these tasks within the guidelines of the PA program but also to serve as skillful mentors that stimulate reflective practice and continued growth. A Peer Assistance Committee, jointly designed by the local association and district, should provide the following governance duties:

• Develop program parameters (to in-clude but not exclusive to)

o Roles and responsibilities of peer assistants

o Roles and responsibilities of teach-ers receiving coaching

o Protocol for matching peer coaches with mentees (and creating a proto-col for changing matches if neces-sary)

o Clarifying and creating culturally competent equity standards for the program

o Schedule of supporto Articulated activities and taskso Firewall protocolso Alignment to evaluation system and

professional learning activitieso Roles and responsibilities of admin-

istrators and other staff in support-ing this program

o Recruiting criteria and training for peer coaches

o Identification of who receives peer coaching and when

• Recruit, train and evaluate peer as-sistants

• Oversee the program budget• Facilitate ongoing evaluation and im-

provement of program

PEER ASSISTANCE

Evaluation ProceduresVIII.

Page 51: OEA's Teacher Evaluation & Support System Guidebook

51

Safeguards should exist to do the following:

Peer coaches do not replace the duties of a licensed administrator who is responsible for summative evaluation

Peer coaches do not provide information or testimony that is used in a summative evaluation

Peer coaches do not breech confidentiality guidelines with other staff

VIII.

The following guiding principles for any Peer Assistance Program are recom-mended in the NEA Teacher Evaluation and Accountability Toolkit:

• Peer Assistance Programs must be col-laboratively developed and overseen by joint labor-management committees

• Before any program is implemented the local association must educate members and build support for it. Any acceptance of this program as part of a collective bargaining agreement hinges on whether or not members believe the PA program is fair, their rights are protected and it has value for them as individuals and a collective.

• Peer coaches must be carefully selected for their skills and commitment to the program.

• Peer coaches who also serve other du-ties must have adequate time to devote to the preparation and implementation of their duties as peer coaches.

• Peer coaches remain bargaining unit members and do not cross over into summative evaluation.

• Confidentiality and firewall protocols are carefully and clearly defined in the

joint committee rules as well as the col-lective bargaining agreement.

• Additional compensation for peer coaches needs to be significant enough, reflecting the additional responsibili-ties, learning, collaboration and time involved. This compensation should be in addition to continued compensation packages and steps owed to them.

The school district should provide a sustainable, stable funding source for the program. A data tracking system should be devel-oped so that the joint PA committee can analyze, modify and improve the program accordingly. Samples of data that team may be interested in are:

• Number of teachers served by the program (divided by probationary and career teachers)

• Financial data• Caseload data• Time served by peer assistants• Retention over time• # of Plans of Assistance over time

Page 52: OEA's Teacher Evaluation & Support System Guidebook

52

A variety of approaches to teacher observation support professional growth and student achievement. The following are several of those methods:LESSON STUDYIn this three-pronged approach designed by Japanese educa-tors, teachers collaboratively develop a lesson, observe it being taught to students, and then discuss and refine it.

PEER COACHINGIn this non-evaluative professional development strategy, edu-cators work together to discuss and share teaching practices, observe each other's classrooms, provide mutual support, and, in the end, enhance teaching to enrich student learning.

COGNITIVE COACHINGTeachers are taught specific skills that involve asking ques-tions so that the teacher observed is given the opportunity to process learning associated with teaching the lesson.

CRITICAL FRIENDS GROUP (CFG)This program provides time and structure in a teacher's sched-ule for professional growth linked to student learning. Each CFG is composed of eight to 12 teachers and administrators, under the guidance of at least one coach, who meet regularly to develop collaborative skills, reflect on their teaching practices, and look at student work. [See an Education World article, Critical Friends Groups: Catalysts for School Change.]

LEARNING WALKThe Learning Walk, created by the Institute for Learning at the University of Pittsburgh, is a process that invites par-ticipants to visit several classrooms to look at student work and classroom artifacts and to talk with students and teach-ers. Participants then review what they have learned in the classroom by making factual statements and posing questions about the observations. The end result is that teachers become more reflective about their teaching practices. Professional development is always linked to The Learning Walks.

From “Teachers Observing Teachers: A Professional Develop-ment Tool for Every School”

Evaluation ProceduresVIII.

52

Page 53: OEA's Teacher Evaluation & Support System Guidebook

53

RELEVANT EXAMPLES OF BARGAINING LANGUAGE, RESEARCH AND OTHER RESOURCES:

• New York: Rochester Teachers Asso-ciation, Section 53, Intervention, Re-mediation, and Professional Support

o Bargaining language for working with career teachers: Page 142 of the NEA Teacher Evaluation and Ac-countability Toolkit

• New Jersey: Support on Site summa-rizes the New Jersey Education Asso-ciation’s induction program

o Page 108 & 109 of the NEA Teacher Evaluation and Accountability Toolkit

• Supporting Research on both Peer Assistance and Peer Assistance and Review Programs

o Page 21 of the NEA Teacher Evalua-tion and Accountability Toolkit

• Boston Teachers Union (www.btu.org/member-resources )

o Website advertises both the reasons why you might want a peer coach and the benefits you might realize in the following way:

PEER ASSISTANCE PROGRAM: WE’RE HERE TO HELP

• Are you overwhelmed by the district’s mandates related to your content area or instructional implementation?

• Have you recently received administra-tive feedback that was a surprise to you?

• Are you looking to refine or polish your teaching craft? Do you need support?

PEER ASSISTANCE PROGRAM: ABOUT OUR WORK

We realize that each teacher’s situation is different. While all our work is confi-dential and non-evaluative, together we might:• Focus on the Eight Dimensions of Ef-

fective Teaching• Reflect to increase student engagement

and learning• Find and create teaching resources and

materials• Collaborate with other teachers and

school leaders.

• On-site support is dictated by your in-dividualized needs. Together we will de-cide the length and focus of our work.

PA (PEER ASSISTANCE) VERSUS PAR (PEER ASSISTANCE REVIEW)Peer Assistance (PA) programs are typi-cally joint, collaborative labor-manage-ment programs that focus on enhancing teacher quality by using a cadre of experi-enced and expert teachers as mentors or coaches for probationary and/or career teachers in their formative assessment. Peer Assistance Review (PAR) programs are similar except that cadre of coaches is also involved in the summative evaluation process of other teachers.

CD-ROM Resources

Evaluation ProceduresVIII.

See the CD-ROM for a longer resource on how in Oregon, only PA is allowed per Oregon Statutes, Or-egon Administrative Rules, and Oregon Public Employ-ee Collective Bargaining Agreement.

Teacher Evaluation & Support System Guidebook

Page 54: OEA's Teacher Evaluation & Support System Guidebook

Questions about this Guidebook?

CONTACT Center for Great Public Schools

[email protected]

Oregon Education Association6900 SW Atlanta St.Portland, OR 97223

www.oregoned.org