Wisconsin Educator Effectiveness System Developmental Pilot

Transcription

Wisconsin Educator Effectiveness System Developmental Pilot
Wisconsin Educator Effectiveness System
Developmental Pilot
Process Manual for Teacher Evaluation of Practice
DRAFT
June 27, 2012
1
Dear Wisconsin Education Leaders:
The great state of Wisconsin has been recognized as an educational leader in
this nation for some time as our work has always been focused around
student achievement. It’s no surprise that as a nation, educational reform
initiatives are directed at teacher quality as it is the most significant factor
impacting student achievement.
Wisconsin’s record of educational
achievement is the result of the hard work of so many quality educators and
education leaders to improve outcomes for all students.
For the past two years, the state has been working to develop a new
evaluation system for teachers and principals – a system that focuses on
professional growth and student achievement. Together, stakeholder groups
representing teachers, administrators, school boards, higher education, and
others have developed the Wisconsin Educator Effectiveness System which
is grounded on the 2011 Interstate Teacher Assessment and Support
Consortium (InTASC) Model Core Teaching Standards and—for
principals—the 2008 Interstate School Leaders Licensure Consortium
(ISLLC) Educational Leadership Policy Standards. The implementation of
this evaluation system represents the best thinking of stakeholders statewide.
This Wisconsin Educator Effectiveness Process Guide is an essential tool to
ensure the success of this effort. The intent of the Educator Effectiveness
System is to support educators in professional development for the benefit of
our students. In addition to this process guide, training and ongoing
professional development will strengthen district implementation as the state
works toward successful statewide implementation of the system in 2014-15.
This work represents a significant change for our educational system, and a
transformation such as this takes time and practice. This gradual
implementation will allow schools to be well supported and successful, and
provides feedback opportunities that will help to inform the refinement of the
components as we near statewide implementation. Continued communication
and feedback from districts and stakeholder groups will help the state of
Wisconsin to develop one of the best evaluation systems in the country.
Sincerely,
Tony Evers, PhD
State Superintendent
2
TABLE OF CONTENTS
Letter from State Superintendent Tony Evers
Developmental Pilot Introduction
Section 1: The Teacher Evaluation Process at a Glance
Section 2: Classroom Practice Measure - Framework for Teaching
Section 3: The Teacher Practice Evaluation Process – Steps, Tasks and Forms
Orientation
Self-Reflection and Educator Effectiveness Plan
Evaluation Planning Session
Observations and Evidence Collection
Pre- and Post-Observation Feedback
Rating of Practice
Final Evaluation Conference
Use of Evaluation Results
Developmental Pilot Evaluation Process Responsibilities
Section 4: Evaluation of Developmental Pilot
Purpose and Basic Design
Participating Districts’ Roles in the Evaluation of the Developmental Pilot
Teacher Evaluation During the Developmental Pilot
Section 5: Definitions
Section 6: Frequently Asked Questions
Appendices
APPENDIX A: Wisconsin Framework for Educator Effectiveness Design Team Report and
Recommendations
APPENDIX B: Danielson 2011 Framework for Teaching Evaluation Rubric
APPENDIX C: DRAFT Wisconsin Teacher Self-Rating Form
APPENDIX D: DRAFT Wisconsin Teacher Educator Effectiveness Plan
3
APPENDIX E: DRAFT Wisconsin Teacher Observation/Artifact Form
APPENDIX F: DRAFT Wisconsin Teacher Pre-Observation (Planning) Form
APPENDIX G: DRAFT Wisconsin Teacher Post-Observation (Reflection) Form
APPENDIX H: DRAFT Wisconsin Teacher Final Evaluation Form
APPENDIX I: Evidence Sources
4
DEVELOPMENTAL PILOT INTRODUCTION
Thank you for participating in the Developmental Pilot of the process for evaluating teaching practice.
Your involvement in this pilot is a critical part of learning how this evaluation process is carried out and
can potentially be improved. The pilot builds on a six-month design process carried out by Wisconsin
educators. The focus this year is only on the teacher professional practice components of the Educator
Effectiveness System which includes the use of a rubric, evidence sources, and a process and timeline for
carrying out evaluation activities. This pilot does not include outcome measures that will be part of the
full Educator Effectiveness System. The outcome measures will be pilot tested in 2013-14 (for more
information, see Wisconsin Framework for Educator Effectiveness Design Team Report and
Recommendations, Appendix A)
This guide is organized into six sections to help evaluators and teachers as they test the process for
evaluating teaching:






Section 1 provides a brief overview of the evaluation process
Section 2 gives an overview of the 2011 Framework for Teaching©, the rubric for teacher
professional practice evaluation
Section 3 provides an in-depth description of the process for teacher evaluation
Section 4 describes the Developmental Pilot
Section 5 lists terminology and definitions of key terms used in this guide
Section 6 addresses answers to frequently asked questions
5
SECTION 1: THE TEACHER EVALUATION PROCESS AT A GLANCE
The Wisconsin Teacher Evaluation system is structured on a performance management cycle. Figure 1
identifies the key components in the cycle.
Orientation
SelfReflection
and Educator
Effectiveness
Plan
Use of
Evaluation
Results
Final
Evaluation
Conference
Evaluation
Planning
Session
Observations
and Evidence
Collection
Rating of
Practice
Pre- & PostObservation
Discussions
Figure 1: Teacher Evaluation Cycle
Orientation: Teacher and evaluator review the evaluation policy and procedures, evaluation rubrics,
timelines, and forms.
Self-Reflection and Educator Effectiveness Plan: Teacher reviews the Framework for Teaching, selfassesses, sets goals, and completes the Educator Effectiveness Plan (EEP).
Evaluation Planning Session: Teacher and evaluator meet to review EEP and set evidence collection
schedule.
Observations and Evidence Collection: Throughout the year, teachers are observed and other evidence is
gathered by teachers and evaluators for the evaluation. Teachers should receive ongoing feedback.
Pre- and Post-Observation Discussions: at least one observation includes a pre-observation discussion and
a post-observation discussion between teacher and evaluator.
Rating of Practice: Throughout the year, evaluator gathers evidence to rate practice according to the
rubric. The ratings should occur after all evidence is collected for the relative components and domains.
6
Final Evaluation Conference: Teacher and evaluator meet to review evidence collected, EEP, progress
made on goals, and Professional Practice Rating.
Use of Evaluation Results: Teachers personalize professional development based on evaluation results.
They set new goals for following year’s EEP.
7
SECTION 2: TEACHING EVALUATION OF PRACTICE BASED ON FRAMEWORK FOR
TEACHING
Data for teacher evaluation will come from classroom observations and other evidence sources. Charlotte
Danielson’s 2011 Framework for Teaching©, a research-based model for assessing and supporting
teaching practice, has been adopted for evaluating teachers. The infrastructure and support for utilizing
Danielson’s Framework will be accessed through Teachscape. Teachscape provides a combination of
software tools, the latest research-based rubrics, electronic scheduling, and observation and data
collection tools necessary for carrying out the evaluation process.
The Framework is organized into 4 domains and 22 components. Evidence can be gathered for all
components, although only domains 2 and 3 are usually observed during a classroom lesson. The full
rubrics can be found in Appendix B. A complete description of the Domains and Components, as well as
indicators and descriptions of performance levels is available in The Framework for Teaching Evaluation
Instrument, 2011 Edition, which participants in the Developmental Pilot will receive. The four
Framework domains are as follows:
 Domain 1: Planning and Preparation
Defines how a teacher organizes the content that the students are to learn—how the teacher designs
instruction. All elements of the instructional design—learning activities, materials, assessments, and
strategies—should be appropriate to both the content and the learners. The components of domain 1
are demonstrated through the plans that teachers prepare to guide their teaching. The plan’s effects
are observable through actions in the classroom.
 Domain 2: The Classroom Environment
Consists of the non-instructional interactions that occur in the classroom. Activities and tasks
establish a respectful classroom environment and a culture for learning. The atmosphere is
businesslike; routines and procedures are handled efficiently. Student behavior is cooperative and
non-disruptive, and the physical environment supports instruction. The components of domain 2 are
demonstrated through classroom interaction and are observable.
 Domain 3: Instruction
Consists of the components that actually engage students in the content. These components represent
distinct elements of instruction. Students are engaged in meaningful work that is important to students
as well as teachers. Like domain 2, the components of domain 3 are demonstrated through teacher
classroom interaction and are observable.
 Domain 4: Professional Responsibilities
Encompasses the teacher’s role outside the classroom. These roles include professional
responsibilities such as self-reflection and professional growth, in addition to contributions made to
the schools, the district, and to the profession as a whole. The components in domain 4 are
demonstrated through teacher interactions with colleagues, families, and the larger community.
Framework for Teaching
8
Domain 1: Planning and Preparation
1a Demonstrating Knowledge of Content and
Pedagogy
1b Demonstrating Knowledge of Students
1c Setting Instructional Outcomes
1d Demonstrating Knowledge of Resources
1e Designing Coherent Instruction
1f Designing Student Assessments
Domain 4: Professional Responsibilities
4a Reflecting on Teaching
4b Maintaining Accurate Records
4c Communicating with Families
4d Participating in a Professional Community
4e Growing and Developing Professionally
4f Showing Professionalism
Domain 2: Classroom Environment
2a Creating an Environment of Respect and
Rapport
2b Establishing a Culture for Learning
2c Managing Classroom Procedures
2d Managing Student Behavior
2e Organizing Physical Space
Domain 3: Instruction
3a Communicating With Students
3b Using Questioning and Discussion Techniques
3c Engaging Students in Learning
3d Using Assessment in Instruction
3e Demonstrating Flexibility and Responsiveness
The Framework for Teaching© defines four levels of performance for each component. The levels of
performance describe teaching practice for a specific lesson (not the teacher). The levels of performance
are defined as follows:
Highly Effective
(Level 4)
Effective
(Level 3)
Minimally Effective
(Level 2)
Ineffective
(Level 1)
Refers to teaching that
does not convey
understanding of the
concepts underlying the
component. This level of
performance is doing harm
in the classroom.
Refers to teaching that has
the necessary knowledge
and skills to be effective,
but its application is
inconsistent (perhaps due
to recently entering the
profession or recently
transitioning to a new
curriculum, grade level, or
subject).
Refers to successful,
professional practice. The
teacher consistently
teaches at a proficient
level. It would be expected
that most experienced
teachers would frequently
perform at this level.
Refers to professional
teaching that innovatively
involves students in the
learning process and
creates a true community
of learners. Teachers
performing at this level are
master teachers and
leaders in the field, both
inside and outside of their
school.
Evidence of teaching practice relative to the Framework for Teaching components will be gathered
through classroom observations, artifacts such as student work samples and logs of parent
communications, and conversations about practice with the evaluator. Teachers will typically demonstrate
varying degrees of proficiency on the different components. This variation is expected. Perfection may
9
be the ideal, but no one can perform at the highest levels all of the time. New teachers may perform at the
basic level most of the time while working toward proficiency. Experienced teachers should be practicing
at the proficient level for most components most of the time. Teachers may be at the distinguished level
on some components, while demonstrating proficiency in other areas.
The following is an example of the rating rubric with descriptions of levels of performance pertaining to
component 1a: Knowledge of Content and Pedagogy, which falls under the domain of Planning and
Preparation (see Appendix B for full rubric).
1a. Knowledge of Content and Pedagogy
Ineffective
In planning and practice,
teacher makes content
errors or does not correct
errors made by students.
Teacher’s plans and
practice show little
understanding of
prerequisite relationships
important to student’s
learning of the content.
Teacher shows little or no
understanding of the range
of pedagogical approaches
suitable to student’s
learning of the content.
Minimally Effective
Teacher is familiar with
the important concepts in
the discipline but displays
lack of awareness of how
these concepts relate to
one another.
Teacher’s plans and
practice indicate some
knowledge of prerequisite
relationships, although
such knowledge may be
inaccurate or incomplete.
Teacher’s plans and
practice reveal a limited
range of pedagogical
approaches to the
discipline or to the
students.
Effective
Teacher displays solid
knowledge of the
important concepts of the
discipline and the way
they relate to one another.
Teacher’s plans and
practice reflect accurate
knowledge of prerequisite
relationships among topics
and concepts.
Teacher’s plans and
practice reflect familiarity
with a wide range of
pedagogical approaches in
the discipline.
Highly Effective
Teacher displays extensive
knowledge of the
important concepts of the
discipline and the ways
they relate both to one
another and to other
disciplines.
Teacher’s plans and
practice reflect knowledge
of prerequisite
relationships among topics
and concepts and provide a
link to necessary cognitive
structures needed by
students to ensure
understanding.
Teacher’s plans and
practice reflect familiarity
with a wide range of
pedagogical approaches in
the discipline, anticipating
student misconceptions.
10
SECTION 3: THE TEACHER PRACTICE EVALUATION PROCESS – STEPS, TASKS AND
FORMS
The following section details the eight steps in the process for Teacher Evaluation, including descriptions
of each task, timing, teacher and evaluator responsibilities and necessary forms used during the
Developmental Pilot.
Step 1: Orientation – Summer of 2012
The orientation process for the Developmental Pilot starts with training for teachers and evaluators. The
Department of Public Instruction hosts this training for three full days during the summer of 2012. In
addition to covering the expectations and formative aspects of the Developmental Pilot, the training
familiarizes participants with Danielson’s 2011 Framework for Teaching© and its use in evaluations of
teacher practice. The training also includes instruction on how to use Teachscape© software for
communication, data collection, and accessing professional resources. All teacher evaluators (typically
principals) participating in the Developmental Pilot receive the training but have additional on-line
training. The on-line training allows evaluators do practice evaluations at any time. It includes a
certification assessment so that observers will be able to match expert evaluators before using the 2011
Framework in “real-time.”
Step 2: Self-Reflection and Educator Effectiveness Plan - September
Each teacher participating in the Pilot will self-assess his or her practice during September, by reflecting
on practice and completing the Wisconsin Teacher Self-Rating Form (see Appendix C).
The self-reflection informs the development of the Educator Effectiveness Plan (EEP). The selfreflection process begins with a review of the Framework for Teaching© and the completion of the SelfRating Form. After self-rating, the teacher creates two to three professional growth goals related to
Framework components that were identified as areas for development. The two to three professional
growth goals guide the evaluation activities for the year, but all of the components from the Framework
for Teaching© will be assessed. The teacher then incorporates these professional growth goals into the
creation of the EEP. The teacher includes the following information on the EEP (see Appendix D):
1.
2.
3.
4.
5.
Two to three goals related to practice1
Domain related to the goal
Actions needed to meet the goals, including how the goals will be measured
Timelines for the completion of the goals
What evidence will be collected to demonstrate progress
Teachers should identify SMART goals: SMART goals are Specific, Measureable, Attainable, Realistic
and Timely. These goals should be relevant to current practice and classroom needs.
1
During the Developmental Pilot, the goals should relate to specific components of the 2011 Framework for
Teaching. In the future, the goals may also relate to school improvement goal(s); Professional Development Plan
(PDP) goal(s); team goal(s) and/or Student Learning Objective (SLO) goal(s).
11
The teacher submits the completed EEP and the Self-Rating Form to the evaluator prior to the
Evaluation Planning Session.
Step 3: Evaluation Planning Session – September or October
Also during September or October, the teacher meets with his or her evaluator in an Evaluation Planning
Session. At this session, the teacher and evaluator collaborate to complete the following activities:
1. Review and agree upon the goals, actions, resource needs and measures of goal completion to
meet the EEP
2. Set the evaluation schedule including scheduled observations and meetings and/or methods of
collecting other sources of evidence (see Appendix I for descriptions of other sources)
Step 4: Observations, Evidence Collection & On-going Feedback – October through March
Observations and Evidence Collection take place from October through March. The teacher is observed
by his or her evaluator2. The following are the minimum observations to be conducted:
Frequency
1 announced observation
1 unannounced observation
3-5 informal and unannounced observations
(walkthroughs)
Duration
45 minutes or (2) 20-minute observations
45 minutes or (2) 20-minute observations
At least 5 minutes
Other evidence shall be provided by the teacher as agreed upon in the Evaluation Planning Session,
such as:
 Lesson/unit plans
 Classroom artifact
o Teacher-created assessments
o Grade Book
o Syllabus
 Communications with families
o Logs of phone calls/parent contacts/emails
o Classroom newsletters
 Logs of professional development activities
Verbal and/or written feedback should follow each observation.
In addition to observations, additional evidence identified during the Evaluation Planning Session is
collected and shared during this period, as agreed upon during the Evaluation Planning Session. The
2
For purposes of the pilot, the administrator will be the primary observer but a peer mentor/evaluator may also
be implemented. In future years, observations may be done by a combination of administrator and peer mentor.
12
Observation/Artifact Form (see Appendix E)3 is used to document evidence gathered pertaining to a
teacher’s professional practice.
Step 5: Pre- & Post-Observation Discussion - October through March
Discussions between teachers and evaluators take place throughout the year, and can be formal
conferences or informal means of delivering feedback (written or verbal). At least one of the observations
must include both a pre- and a post-observation discussion, which involves the use of the PreObservation (Planning) Form and the Post-Observation (Reflection) Form.
Teachers complete the Pre-Observation (Planning) Form (see Appendix F) in advance of the preobservation discussion. This form helps shape the dialog of the pre-observation discussion and allows the
teacher to “set the stage” for the lesson. The information allows the teacher to identify the context of the
classroom, the specifics of the lesson focus and its intended outcomes. The teacher may submit this
information in writing or come prepared to have a dialog with the evaluator during the pre-observation
discussion.4
The Post-Observation (Reflection) Form (see Appendix G) also helps frame the dialog and resulting
feedback from the observed lesson during the post-observation discussion. Both the teacher and evaluator
can use the questions to identify areas of strength and suggestions for improvement. The post-observation
discussion can also focus on classroom teaching artifacts (lesson plans, student work samples, etc.) that
are related to the classroom observation.
Both the pre-and post-observation forms and discussions can also help provide evidence of reflective
skills that are relevant to Domain 4 of the Danielson Framework for Teaching©.
Step 6: Rating of Practice – February through end of school year
Throughout the school year, evaluators collect and teachers provide evidence of teaching practice. This
evidence is used to rate a teacher’s practice, using the rubric and appropriate levels of performance.
Although evidence is collected throughout the year, evaluators should not make ratings of practice until
adequate information is obtained to assess each component. This will likely occur during the second half
of the school year. Evaluators will need to organize the observation/artifact forms used to record
evidence and any notes compiled during the year (e.g., notes from the post-observation discussions and/or
other feedback sessions). Once the evaluator determines that there is enough evidence for each
component, they will select the rubric level that best matches the evidence of practice for that component.
Step 7: Final Evaluation and Final Evaluation Conference – April - June
3
Observations and artifact collection using Teachscape will be accessed through the Teachscape ReflectLive
feature.
4
These forms will also be accessible using Teachscape ReflectLive.
13
The Final Evaluation Conference takes place during April, May or June. During this conference, the
teacher and his/her supervising administrator meet to discuss achievement of the goals identified in the
EEP and review collected evidence. The teacher submits the completed EEP prior to the Final Evaluation
Conference. The administrator then provides written feedback for the goals and components identified in
the EEP. In addition to the EEP, other collected evidence is used by the evaluator to rate each of the
components from the four domains and twenty-two components. The evaluator then completes the Final
Evaluation Form (Appendix H) and reviews it with the teacher during the Final Evaluation Conference.
The teacher has the opportunity to comment on the final evaluation results on the Final Evaluation
Form.
Step 8: Use of Evaluation Results – end of evaluation year, beginning of following year
Results from the evaluation process inform performance goals and the teacher’s Educator Effectiveness
Plan for the following year. During the Developmental Pilot, NO evaluation results will be used for
employment purposes or other high-stakes decision-making.
14
Developmental Pilot Evaluation Process Responsibilities
The following lists the responsibilities for Teachers and for the Supervising Administrator during the
evaluation process.
Teacher responsibilities












Actively participate in evaluation activities.
Attend the evaluation Orientation.
Reflect on practice and, using the 2011 Framework for Teaching©, complete the Self-rating of
performance.
Based on the Self-rating of performance, identify 2-3 goals related to practice to include in an
Educator Effectiveness Plan (EEP).
Submit the Self-rating of performance and the EEP to the evaluator prior to the Evaluation
Planning Session.
Meet with evaluator for the Evaluation Planning Session, discuss actions needed to achieve
goals identified in EEP, agree upon measures of goal completion, determine evidence sources;
schedule observations and outline timing/process for other evidence collection.
For formal observations be prepared for pre-observation and post-observation meetings. The
teacher may submit this information in writing or come prepared to discuss the information.
Work collaboratively with the peer mentor or secondary observer (if applicable).
Provide the supervising administrator with other evidence as appropriate prior to Final
Evaluation.
Prepare for the Final Evaluation and provide any evidence requested by the supervising
administrator.
Meet with the supervising administrator for the Final Evaluation Conference, review EEP and
goal achievement; sign off on Professional Practice Rating.
Use Evaluation results to inform performance goals and professional development planning for
the following year.
Supervising Administrator responsibilities:










Attend training sessions and be certified in 2011 Framework for Teaching© evaluation.
Participate in the Orientation.
Review EEP prior to the Evaluation Planning Session.
Schedule and facilitate the Evaluation Planning Session, discuss actions needed to achieve goals
identified in EEP, agree upon measures of goal completion; determine evidence sources;
schedule observations and outline timing/process for other evidence collection.
Complete a minimum of one announced observation of 45 minutes or two 20-minute
observations.
Complete a minimum of one pre-conference and post-conference observation with the teacher.
Complete one unannounced observation of 45 minutes or two 20-minute observations.
Complete 3-5 informal and unannounced observations (walkthroughs), of at least 5 minutes.
Provide constructive, written or verbal formative feedback within one week of the observations.
Monitor and review evidence collection throughout the year.
15


Prepare for and schedule the Final Evaluation Conference, review EEP and evidence collected,
assign rating level for each component, and determine Professional Practice Rating.
Facilitate the Final Evaluation Conference using the Final Evaluation form; review goal
achievement, provide formative feedback, identify growth areas for the following year, and
provide Professional Practice Rating.
16
SECTION 4: EVALUATION OF DEVELOPMENTAL PILOT
Purpose and Basic Design
During the 2012-13 school year, the Developmental Pilot of the Wisconsin Educator Effectiveness
System will be evaluated by researchers from the Wisconsin Center for Education Research (WCER) at
UW-Madison. The purpose of the evaluation is to study the quality of implementation of certain aspects
of the Educator Effectiveness System and to provide responsive feedback to stakeholders as these aspects
are tested by schools and districts. Parts of the evaluation will include analysis of implementation data,
observations, surveys of teachers and principals, and interviews and focus groups of participating
teachers, principals, and others. Topic areas will include understanding, attitudes, training received and
needed, data utilization, and feedback on the pilot implementation.
Since parts of the Educator Effectiveness System are still in development, the evaluation will not yet
focus on impact on teacher effectiveness or student learning. The success of the evaluation will be
determined by the pilot’s use of information obtained for improving the execution of the system to match
its intended design. In subsequent years, the evaluation will gradually change to focus more on impact as
the system’s aspects become more routinized.
Participating Districts’ Roles in the Evaluation of the Developmental Pilot
Districts and the individual educators participating in the Developmental Pilot are expected to help
facilitate access to the data necessary for evaluating the Pilot. Data collection may include participating in
individual or group interviews (focus groups), completing surveys, and/or sharing documents relevant to
the system as requested by DPI and the evaluation team at WCER. This feedback will be critical as DPI
works to learn from educators in the pilot and make needed adjustments in the System.
Since the Educator Effectiveness System is still in development, any outcomes, including educator
evaluation ratings, should NOT be used for employment or other high-stakes decisions. If an educator is
identified during the Pilot as having serious performance deficiencies, that individual should be removed
from participation in the pilot study and the district should implement their existing personnel support
structures to address the situation.
Teacher Evaluation During the Developmental Pilot
Teachers will be evaluated by their supervising administrator (the principal or a designee). Those districts
piloting the teacher practice evaluation process will include a small number of teachers (about 3) and their
evaluators (principal or associate principal) and peer evaluators.
17
SECTION 5: DEFINITIONS
The following are definitions for terminology relevant to the teacher practice evaluation system during the
Developmental Pilot.
Announced observation – a formal and pre-scheduled observation. It may be preceded by a preobservation discussion and followed by a post-observation discussion where verbal and/or written
feedback is provided by the evaluator to the teacher.
Artifacts – forms of evidence that support an educator’s evaluation. They may include lesson plans,
examples of student work with teacher feedback, professional development plans and logs of contacts
with families. Artifacts may take forms other than documents, such as videos of practice, portfolios, or
other forms of evidence.
Components - the descriptions of the aspects of a domain, there are 22 components in the 2011
Danielson Framework for Teaching©.
Developmental Pilot – the limited test of a system to further refine its components and processes. The
Wisconsin Model Educator Effectiveness System is undergoing a Developmental Pilot in 2012-2013 in
approximately 100 volunteer districts across the state.
Domains - There are four domains, or broad areas of teaching responsibility, included in the 2011
Framework for Teaching©: Planning & Preparation, Classroom Environment, Instruction, and
Professional Responsibilities. Under each domain, 5-6 components describe the distinct aspects of a
domain.
Educator Effectiveness System– the Wisconsin state model for teacher and principal evaluation, built by
and for Wisconsin educators. Its primary purpose is to support a system of continuous improvement of
educator practice—from pre-service to inservice—that leads to improved student learning. The Educator
Effectiveness System is legislatively mandated by 2011 Wisconsin Act 166.
Educator Effectiveness Plan (EEP) – a document that lists the specific year-long goals for an educator,
along with the actions, resource needs required to attain these goals and the measures necessary to
evaluate the progress made on them. The EEP also allows for a review of progress made on each goal at
the end of the year and serves as the organizer for feedback given by the educator’s evaluator during the
Final Evaluation Conference.
Evaluation Planning Session – A conference in the fall during which the teacher and his or her primary
evaluator discuss the teacher’s Self-Rating and Educator Effectiveness Plan, agree upon goals and actions
needed to meet goals, set an evaluation schedule and schedule any announced observations and the timing
and process for other evidence collection.
Evaluation Rubric – an evidence-based set of criteria across different domains of professional practice
that guide an evaluation. Practice is rated across four rating categories that differentiate effectiveness,
with each rating tied to specific look-fors to support the ratings.
18
Evidence collection – the systematic gathering of evidence that informs an educator’s practice
evaluation. In the Educator Effectiveness System, multiple forms of evidence are required to support an
educator’s evaluation and are listed in this guide in Appendix D.
Final Evaluation Conference – the teacher and his/her evaluator meet to discuss achievement of the
goals identified in the Educator Effectiveness Plan, review collected evidence, and discuss results and
ratings.
Formative Evaluation – the systematic gathering of information with the purpose of understanding an
educator’s strengths and weaknesses in order to improve teaching and learning.
Framework – the combination of the evaluation rubric, the collection and rating of evidence sources, and
the process for evaluating an educator. The Framework is the basis for the evaluation process and the
determination of an overall performance rating.
Goal setting /Professional growth goals – establishing practice related goals are an important part of
professional practice. Goals are set as educators prepare for their Educator Effectiveness Plan, and they
are monitored by the educator during the year.
Indicators/Look-fors – observable pieces of information for evaluators to identify or “look-for” during
an observation or other evidence gathering. Indicators are listed in the Sources of Evidence (Appendix D).
Inter-Rater Agreement – the extent to which two or more evaluators agree in their independent ratings
of educators’ effectiveness.
Observations – one source of evidence informing the evaluation of a teacher. Observations may be
announced (scheduled in advance, possibly with a pre- and/or post-observation conference) or
unannounced; formal (lengthy and with conferences) or informal (short and impromptu). Observations are
carried out by the educator’s evaluator, who looks for evidence in one or more of the components of the
Framework for Teaching© evaluation rubric.
Orientation – The first step in the Educator Effectiveness evaluation process, the Orientation takes place
prior to or at the beginning of the school year. Educators receive training in the use of their professional
practice frameworks, the related tools and resources, timelines for implementation, and expectations for
all participants in the system.
Peer evaluator - A teacher who is trained to do informal observations. A peer evaluator may be a
department-head or grade-level lead or other instructional leader. Peer evaluators are trained in the
Framework and Educator Effectiveness process.
Post-observation conference - a conference that takes place after a formal observation during which the
evaluator provides feedback verbally and in writing to the teacher.
Pre-observation conference - a conference that takes place before a formal observation during which the
evaluator and teacher discuss important elements of the lesson or class that might be relevant to the
observation.
19
Student Learning Objectives (SLOs) - a specific learning outcome measuring student achievement
and/or growth throughout the year. More details on Student Learning Objectives can be found in the
Student Learning Objectives process manual.
Self-rating of performance – teachers will complete a self-assessment at the beginning of the year and
will review it prior to each conference. This self-assessment will ask educators to reflect on their past
performance, relevant student learning data, prior evaluation data, and professional goals for the
upcoming year.
Unannounced Observation – an observation that is not scheduled in advance. No pre-observation
conference is held with an unannounced observation, but written or verbal feedback is expected within
seven days.
Walkthrough - a short (5 minute minimum) informal and unannounced observation of a teacher’s
practice in the classroom.
20
SECTION 6: FREQUENTLY ASKED QUESTIONS
The following are some frequently asked questions specific to the teacher practice evaluation system
during the Developmental Pilot. For a more complete listing of FAQs, please see [Web link]
1) What is the Educator Effectiveness System?
In 2011 the Wisconsin Educator Effectiveness System was endorsed within Wisconsin Act 166. The
purpose is to provide a fair, valid and reliable state evaluation model for teachers and principals that
supports continuous improvement of educator practice resulting in improved student learning.
2) How was the Educator Effectiveness System developed?
State Superintendent Tony Evers convened an Educator Effectiveness System Development Team
that included a broad representation of educators and education stakeholders to develop a framework
for teacher and principal evaluation. A full list of participating organizations is available on the DPI
website. The Design Team’s main recommendations were embraced in Act 166.
3) How will the Educator Effectiveness System be implemented?
The Educator Effectiveness System will be implemented in 3 stages over 4 years:
Stage 1: 2011-12: Evaluation process developed for a) teacher evaluation of practice; b) teacher
developed Student Learning Objectives; c) principal evaluation of practice
Stage 2: 2012-13: Developmental Pilot test of the 3 evaluation processes in school districts that
have volunteered to participate and provide feedback for improvement. Principal School
Learning Objectives will be tested in the spring of 2013.
2013-14: Full Pilot test of the Educator Effectiveness models for Teacher Effectiveness
and Principal Effectiveness. These models will include the measures of practice and
measures of student outcomes (to be developed in 2012-2013). This full pilot will occur in a
sample of schools in all districts.
Stage 3: 2014-15: Educator Effectiveness System implemented state-wide. The system will be
fully implemented statewide, but will still be the subject on an external evaluation to
address the impact and need for any additional improvements.
4) Have other states done this?
Yes, many states have developed or are in the process of developing new teacher and principal
evaluation systems that include both professional practice and student outcome measures. A number
of states have passed legislation requiring the implementation of a statewide system and are on a
timeline similar to Wisconsin.
5) How will staff understand the Educator Effectiveness System?
All teachers participating in the Pilot will receive the same comprehensive training as administrators
and evaluators. A series of informational webinars will be released to participating districts
throughout the pilot years for use by all stakeholder groups.
21
The Framework for Educator Effectiveness is available and the process manuals for the Pilot Phase
will be available on the DPI Educator Effectiveness website.
6) Who will be evaluated using the new Educator Effectiveness System?
In 2012-13 principals and teachers will be evaluated. During 2012-13, teachers include PK – 12 (prekindergarten through grade 12) core content/grade level teachers and teachers of music, art and
physical education. Assistant/associate principals and teacher specialist positions will be added later,
once the evaluation processes are adapted for those educators.
7) Who will evaluate teachers?
During the pilot stage (2012-14), an administrator will be the primary observer. A peer
mentor/observer may also be utilized if agreed to by the teacher. Once the Educator Effectiveness
System is fully implemented, observations may be made by a combination of administrators and peer
mentors/observers.
8) How will evaluators be trained?
Prior to the implementation of an Educator Effectiveness System, evaluators (including peer mentors
and observers involved in this process) will be required to complete a comprehensive certification
training program approved by DPI. Certification will include training on bias and consistency (i.e.,
inter-rater agreement).
9)
Will the Educator Effectiveness System be used to make employment and/or compensation
decisions?
Not yet. During the pilot stage, 2012-14, pilot participants will be evaluated for formative purposes.
Data collected during the pilot stage (2012-14) will not be used to make employment decisions
because the system is still in development. Once the system is fully developed and tested, districts
may choose to apply the systems to human resource decisions, including targeted professional
development. Any performance ratings will not be made public.
10) Do walkthroughs count as observations?
Yes. The process manuals call for a series of observations which include formal, as well as informal
and unannounced observations such as walkthroughs.
11) Does every observation need to have a pre- and post- conference?
No. Pre- and post- conferences are required for at least one formal observation. Teachers may request
additional pre/post conferences with the evaluator or with a peer mentor/observer.
12) What will the evaluation cycle look like?
22
During the Pilot stage, 2012-14, evaluation activities will be used for formative purposes only. Once
fully implemented in 2014-15, the Educator Effectiveness System requires annual evaluations, with a
summative evaluation once in a three year cycle. The other two years will be formative, unless there
is a determination that performance requires more comprehensive evaluation.
13) How will this evaluation system affect a teacher’s daily work load?
The Educator Effectiveness cycle (outlined in the Process Manual for Teacher Evaluation) calls for
Orientation, an Evaluation Planning Session, a series of observations, and a Final Evaluation
conference with the evaluator. These events represent basic performance management processes that
are common in high performing organizations. They will be coordinated by the teacher and the
evaluator/mentor/observer. Teachers will also be required to collect classroom artifacts as evidence of
ongoing effectiveness. This process will be addressed during training for the Educator Effectiveness
System.
14) What if a teacher does not agree with the evaluation?
During the developmental pilot, there will be no high stakes decisions connected with the evaluation
process. If a teacher disagrees with their evaluation of practice, s/he can comment about the
disagreement on the final evaluation form. Most districts already have an appeals process for
personnel evaluations, but if not, they will need to develop an appeals process once the Educator
Effectiveness System is fully implemented.
15) How will teachers and others provide feedback on the pilot experience?
An external study will be conducted with pilot districts to collect ongoing feedback through
interviews, surveys and document analysis. The data collected will provide ongoing feedback which
will be used by the DPI Educator Effectiveness work groups to make improvements to the system. A
designated email will be set up for on-going feedback to DPI on the Educator Effectiveness System,
though this email is not intended to be used for responses to questions or technical support.
16) Based on the results of the pilot programs, will changes be made?
Yes. The Pilot stage is intended to improve the Educator Effectiveness System and meet its intended
purpose: to provide a fair, valid and reliable state evaluation model for teachers and principals that
supports continuous improvement of educator practice resulting in improved student learning.
17) Will all districts be required to implement the same system, or is there flexibility?
There is some flexibility. This system is intended to be fair and consistent across the state. While each
district will have some flexibility in choosing standards and rubrics to assess teacher and principal
professional practice, there will be similar features required to assess student and school outcomes
(e.g., use of Student Learning Objectives and Value –Added Measures of achievement). All models
must be approved through an equivalency process that will be defined by DPI.
23
18) Who should a district contact for technical advice?
There will be designated contacts at DPI for questions about the developmental pilot. Technical
assistance questions can be presented to the DPI contact.
24
APPENDIX A: Wisconsin Framework for Educator Effectiveness Design Team Report and
Recommendations
25
APPENDIX B: Danielson 2011 Framework for Teaching Evaluation Rubric (to come)
26
APPENDIX C: DRAFT Wisconsin Teacher Self-Rating Form
The self-rating process allows teachers to reflect on their practice and prior evaluations and prepare for the development of their Educator Effectiveness Plan.
Using the Wisconsin Teaching Effectiveness Rubric, review the components within each domain and rate yourself accordingly. Based on that rating, identify an
area of strength or an area for development related to that component.
Domain 1
1.a
1.b
1.c
1.d
1.e
1.f
Planning and Preparation
Demonstrating Knowledge of Content and Pedagogy
Demonstrating Knowledge of Students
Setting Instructional Outcomes
Demonstrating Knowledge of Resources
Designing Coherent Instruction
Designing Student Assessments
Ineffective
Minimally Effective
Effective
Highly Effective
Effective
Highly Effective
Effective
Highly Effective
Domain 1: Planning and Preparation - Strength or Area for Development
Based on the above ratings, identify a strength or area for development:
Why did you make this assessment (what evidence was used to make the assessment)?
Domain 2
2.a
The Classroom Environment
Creating Environment of Respect and Rapport
2.b
Classroom Observations & Feedback
2.c
2.d
Managing Classroom Procedures
Managing Student Behavior
Ineffective
Minimally Effective
Domain 2: The Classroom Environment – Strength or Area for Development
Based on above rating, identify a strength or area for development:
Why did you make this assessment (what evidence was used to make the assessment)?
Domain 3
Instruction
Ineffective
Minimally Effective
27
3.a
3.b
3.c
3.d
3.e
Communication with Students
Questioning and Discussion Techniques
Engaging Students in Learning
Using Assessment in Instruction
Demonstrating Flexibility and Responsiveness
Domain 3: Instruction – Strength or Area for Development
Based on above rating, identify a strength or area for development:
Why did you make this assessment (what evidence was used to make the assessment)?
Domain 4
4.a
Professional Responsibilities
Reflecting on Teaching
4.b
Maintaining Accurate Records
4.c
4.d
4.e
4.f
Communicating with Families
Participating in a Professional Community
Growing and Developing Professionally
Showing Professionalism
Ineffective
Minimally Effective
Effective
Highly Effective
Domain 4: Professional Responsibilities – Strength or Area for Development
Based on above rating, identify a strength or area for development:
Why did you make this assessment (what evidence was used to make the assessment)?
28
Additional comments about ratings, strengths, or areas for development:
29
APPENDIX D: DRAFT Teacher Educator Effectiveness Plan
Teacher___________________________________________________ Grade(s)/Subject(s)______________________ Date Reviewed___________________
School_________________________________________________________________________Evaluator__________________________________________
Describe goal:
Identify domain(s) and components related to goal:
Strategies
Timeline
Potential Obstacles
Resources
Evidence Sources
How will you measure progress on goal?
Describe goal:
30
Identify domain(s) and components related to goal:
Strategies
Timeline
Potential Obstacles
Resources
Evidence Sources
How will you measure progress on goal?
Describe goal:
Identify domain(s) and components related to goal:
31
Strategies
Timeline
Potential Obstacles
Resources
Evidence Sources
How will you measure progress on goal?
32
APPENDIX E: DRAFT Wisconsin Teacher Observation/Artifact Form
Teacher_______________________________________________________ Date______________________________
School________________________________________ Observer ___________________________________________
Briefly describe the observation location and activity(s), and/or artifact:
List the relevant component(s) for this observation and/or artifact:
What evidence did you note during the observation and/or from the artifact that applies to the rubric(s)?
Feedback to Teacher:
33
APPENDIX F: DRAFT Wisconsin Teacher Pre-Observation (Planning) Form
1.) To which part of your curriculum does this lesson relate?
2.) How does this learning “fit” in the sequence of learning for this class?
3.) Briefly describe the students in this class, including those with special needs.
4.) What are your learning outcomes for this lesson? What do you want the students to understand?
5.) How will you engage the students in the learning? What will you do? What will the students do? Will the students
work in groups, or individually, or as a large group? Provide any worksheets or other materials the students will be
using.
6.) How will you differentiate instruction for different individuals or groups of students in the class?
7.) How and when will you know whether the students have learned what you intend?
8.) Is there anything that you would like me to specifically observe during the lesson?
APPENDIX G: DRAFT Wisconsin Teacher Post-Observation (Reflection) Form
34
1.) In general, how successful was the lesson? Did the students learn what you intended for them to learn? How do you
know?
2.) If you were able to bring samples of student work, what do those samples reveal about those students’ levels of
engagement and understanding?
3.) Comment on your classroom procedures, student conduct, and your use of physical space. To what extent did these
contribute to student learning?
4.) Did you depart from your plan? If so, how, and why?
5.) Comment on different aspects of your instructional delivery (e.g. activities, grouping of students, materials, and
resources). To what extend were they effective?
6.) If you had a chance to teach this lesson again to the same group of students, what would you do differently?
35
APPENDIX H: DRAFT Wisconsin Teacher Final Evaluation Form
Teacher______________________________________School______________________________________Grade Level_____________________
Evaluator_____________________________________Date__________________________
Domain
Component
Ineffective (1)
1: Planning and
Preparation
1.a Demonstrating Knowledge of Content and
Pedagogy
1.b Demonstrating Knowledge of Students
1.c Setting Instructional Outcomes
1.d Demonstrating Knowledge of Resources
1.e Designing Coherent Instruction
1.f Designing Student Assessments
Evidence/Artifact(s) used for evidence:
Comments:
2: Classroom
Environment
2.a Creating Environment of Respect and
Rapport
2.b Classroom Observations & Feedback
2.c Managing Classroom Procedures
2.d Managing Student Behavior
Evidence/Artifact(s) used for evidence:
Comments:
3.a Communication with Students
3: Instruction
3.b Questioning and Discussion Techniques
3.c Engaging Students in Learning
3.d Using Assessment in Instruction
3.e Demonstrating Flexibility and
Rating
Minimally Effective (2)
Effective (3)
Highly Effective (4)
Responsiveness
Artifact(s) used for evidence:
Comments:
4: Professional
Responsibilities
4.a Reflecting on Teaching
4.b Maintaining Accurate Records
4.c Communicating with Families
4.d Participating in a Professional Community
4.e Growing and Developing Professionally
4.f Showing Professionalism
Artifact(s) used for evidence:
Comments:
Total Ratings for Professional Practice
Overall Professional Practice Rating: _________________________________
Key strengths:
37
Areas for development:
Teacher comments on rating (optional)
Teacher Signature_______________________________________________________________ Date__________________________
Evaluator Signature_______________________________________________________________ Date__________________________
38
APPENDIX I: Teacher Evidence Sources
Component
1a: Demonstrating
knowledge of
content and
pedagogy
1b: Demonstrating
knowledge of
students
Evidence Required or Strongly
encouraged (Bold)
 Evaluator/Teacher
conversations
 Lesson/unit plan
 Observation




Evaluator/Teacher
conversations
Lesson/unit plan
Observation
Student / Parent Perceptions
Domain 1: Planning and Preparation
Indicator/“look-fors” (* = Danielson 2011)
-
Descriptive feedback on student work
Adapting to the students in front of you
Scaffolding based on student response
Questioning process
Teachers using vocabulary of the discipline
Accuracy or inaccuracies in explanations
1.
-
Artifacts that show differentiation
Artifacts of student interests and backgrounds, learning
style, outside of school commitments (work, family
responsibilities, etc.)
Differentiated expectations based on assessment
data/aligned with IEPs
Obtaining information from students about:
background/ interests, relationships/rapport
1.
Same learning target, differentiated pathways
Students can articulate the learning target when asked
Targets reflect clear expectations that are aligned to
standards
Checking on student learning and adjusting future
instruction
Use of entry/exit slips
Diversity of resources (books, websites, articles,
media)
Guest speakers
College courses
Collaboration with colleagues
Connections with community (businesses, non-profit
organizations, social groups)
Evidence of teacher seeking out resources (online or
other people)
1.
-
1c: Setting
instructional
outcomes



Evaluator/Teacher
conversations
Lesson/unit plan
Observation
-
1d: Demonstrating
knowledge of
resources



Evaluator/Teacher
conversations
Lesson/unit plan
Observation
Evidence Collection
-
2.
3.
2.
3.
4.
2.
3.
1.
2.
3.
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection)
Lesson plans/unit plans
Observations
 Notes taken during observation
Optional
 Student / Parent surveys
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation lesson plan
39
1e: Designing
coherent instruction
(use of appropriate
data)






Evaluator/Teacher
conversations
Lesson/unit plan
Observation
Pre observation form
Learning targets
Entry slips/Exit slips
-
Grouping of students
Variety of activities
Variety of instructional strategies
Student voice and choice
Same learning target, differentiated pathways
Linking concepts/outcomes from previous lessons
1.
2.
3.
4.
1f: Designing
student assessment




Evaluator/Teacher
conversations
Lesson/unit plan
Observation
Formative and summative
assessments and tools
-
Component
2a: Creating an
environment of
respect and rapport
Evidence Required or Strongly
encouraged (Bold)
 Evaluator / Teacher
conversations
 Observations
 Video
 Illustrations of response to
student work
Alignment of assessment to learning target
Plan for bringing student assessment into your
instruction (feedback loop from student within
assessment plan)
Uses assessment to differentiate instruction
Expectations are clearly written with descriptions of
each level of performance
Differentiates assessment methods to meet individual
students needs.
Students have weighed in on the rubric or assessment
design
Domain 2: The Classroom Environment
Indicator/”look fors” (* = Danielson 2011)
-
Respectful talk and turn taking
Respect for students’ background and life outside the
classroom*
Teacher and student body language*
Physical proximity*
Warmth and caring*
Politeness*
Encouragement*
Active listening*
Fairness*
Response to student work: Positive reinforcement,
respectful feedback, displaying or using student work
1.
2.
3.
4.
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation
Optional
 Pre observation form
 Learning targets
 Entry / exit slips
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation
Optional
 Formative and Summative assessments and
tools (i.e. rubrics, scoring guides,
checklists)
 Student developed assessments
Evidence Collection
1.
2.
3.
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
 Use questions on observation forms
(especially describing students in class)
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic)
 Observer takes notes during pre- and postobservation conferences
Optional
 Video
 Response to student work
40
2b: Establishing a
culture for learning




Observations
Student Assignments
Lesson plan
Video/Photos
-
Belief in the value of the work*
High expectations, supported through both verbal and
nonverbal behaviors*
Expectation and recognition of quality*
Expectation and recognition of effort and persistence
Confidence in students’ ability evident in teacher’s and
students’ language and behaviors*
Expectation for all students to participate*
Use of variety of modalities
Student Assignments: Rigor, Rubrics Used, Teacher
Feedback, Student Work Samples
Use of Technology: Appropriate Use
1.
2.
3.
2c: Managing
classroom
procedures



Observations
Syllabus
Parent Communication
-
Smooth functioning of all routines*
Little or no loss of instructional time*
Students playing an important role in carrying out the
routines*
Students knowing what to do, where to move*
1.
2.
2d: Managing
student behavior




Observations
Disciplinary records/plans
(content)
Student / Parent Feedback
Parent Communications
-
Clear standards of conduct, possibly posted, and
referred to during a lesson*
Teacher awareness of student conduct*
Preventive action when needed by the teacher*
Fairness*
Absence of misbehavior*
Reinforcement of positive behavior*
Culturally responsive practices
Time on task, posting classroom rules, positive
reinforcement.
1.
2.
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic)
 Observer takes notes during pre- and postobservation conferences
 Observer interacts with student about what
they are learning.
Student Assignments
 Teacher provides examples of student
work.
Optional
 Lesson plan
 Video / Photo
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic)
 Observer takes notes on what is happening
at what time, tracking student engagement
/ time on task, classroom artifacts on
procedures.
Optional
 Syllabus
 Communications to Students / Parents
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic)
 Observer may tally positive reinforcement
vs. punitive disciplinary action.
Optional
 Disciplinary records/plans (content)
 Student / Parent Feedback
 Parent Communications
41
2e: Organizing
physical space
Component
3a: Communicating
with students



Observations
Video/Photos
Online Course Structure
Evidence Required or Strongly
encouraged (Bold)
 Observations
 Assessed Student work
 Communications with students
 Handouts with instructions
 Formative Assessments
-
Pleasant, inviting atmosphere*
Safe environment*
Accessibility for all students
Furniture arrangement suitable for the learning
activities*
Effective use of physical resources, including computer
technology, by both teacher and
Students*
Established traffic patterns
Domain 3: Instruction
Indicator/”look-fors” (* = Danielson 2011)
1.

-
1.
-
3b: Using
questioning and
discussion
techniques





Observations
Lesson Plan
Videos
Student Work
Discussion Forums
-
3c: Engaging
students in learning


Observations
Lesson plans
-
Clarity of the purpose of the lesson*
Clear directions and procedures specific to the lesson
activities*
Absence of content errors and clear explanations of
concepts*
Students comprehension of content*
Correct and imaginative use of language
Assessed student work - specific feedback
Use of electronic communication: Emails, Wiki, Web
pages
Formative Assessments: Exit / Entry Slips

2.
Observation
Observer “scripts” lesson or takes notes on
specially – designed form (paper or electronic).
Observer records classroom physical features
on standard form or makes a physical map.
Optional
 Photos, Videos
 Online course structure
Evidence Collection
2.
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic). Dialogue with students and
accurate / precise dialogue.
 Observer collects examples of written
communications (emails / notes)
Assessed Student Work
 Teacher provides samples of student work
& written analysis after each observation
or end of semester
3.
Optional
 Electronic Communication
 Handouts with instructions
 Formative Assessments
Questions of high cognitive challenge, formulated by
both students and teacher*
Questions with multiple correct answers, or multiple
approaches even when there is a single
correct response*
Effective use of student responses and ideas*
Discussion in which the teacher steps out of the central,
mediating role*
High levels of student participation in discussion*
Student Work: Write/Pair/Share, student generated
discussion questions, online discussion
1.
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic).
 Observer tracks student responses.
2.
Optional
 Lesson Plan
 Videos
 Student Work
 Discussion Forums
Activities aligned with the goals of the lesson*
Student enthusiasm, interest, thinking, problem-
1.
Observations
 Observer “scripts” lesson or takes notes on
42


Student work
Use of technology/instructional
resources
-
3d: Using
assessment in
instruction




Observations
Formative / Summative
Assessment Tools
Lesson plans
Conversations w / Evaluator
-
3e: Demonstrating
flexibility and
responsiveness




Component
4a: Reflecting on
teaching
Observations
Lesson plans
Use of supplemental
instructional resources
Student Feedback
Evidence Required or Strongly
encouraged (Bold)
 Evaluator/Teacher
Conversations
 Observations
 Teacher PD goals/plan
-
solving, etc.*
Learning tasks that require high-level student thinking
and are aligned with lesson objectives*
Students highly motivated to work on all tasks and
persistent even when the tasks are challenging*
Students actively “working,” rather than watching
while their teacher “works”*
Suitable pacing of the lesson: neither dragging nor
rushed, with time for closure and student*
Reflection*
Student – student conversation
Student directed or led activities / content
Teacher paying close attention to evidence of student
understanding*
Teacher posing specifically created questions to elicit
evidence of student understanding*
Teacher circulating to monitor student learning and to
offer feedback*
Students assessing their own work against established
criteria*
Assessment tools: use of rubrics
Formative / Summative assess tools: frequency,
descriptive feedback to students
Lesson plans adjusted based on assessment
Incorporation of student interests and events of the day
into a lesson*
Visible adjustment in the face of student lack of
understanding*
Teacher seizing on a teachable moment*
Lesson Plans: Use of formative assessment, use of
multiple instructional strategies
Domain 4: Professional Responsibilities
Indicator/”look-fors” (* = Danielson 2011)
-
Revisions to lesson plans
Notes to self / journaling
Listening for analysis of what went well and didn’t go
well

specially – designed form (paper or
electronic).
Observer tracks student participation, time
on task, examines student work, and
teacher / student interactions.
2.
Optional
 Lesson plans
 Student work
 Use of technology/instructional resources
1.
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic).
Formative / Summative Assessment Tools
 Teacher provides formative and summative
assessment tools or data
2.
3.
Optional
 Lesson plans
 Conversations w / Evaluator
1.
Observations
 Observer “scripts” lesson or takes notes on
specially – designed form (paper or
electronic).
 Takes notes on teacher taking advantage of
teachable moments
2.
Optional
 Lesson plans
 Use of supplemental instructional
resources
 Student Feedback
Evidence Collection
1.
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
43

Student / Parent Feedback
-
4b: Maintaining
Accurate Records





4c: Communicating
with families


Evaluator/Teacher
Conversations
Lesson/unit plan
Grade Book
Artifact – teacher choice
Systems for data collection
Logs of phone calls/parent
contacts/emails
Observation during parent
teacher meeting or conference
-
-
4d: Participating in a
professional
community




Observation
Attendance at PD sessions
Mentoring other teachers
Seeking mentorship
-
Specific examples of reflection from the lesson
Ability to articulate strengths and weaknesses
Capture student voice (survey, conversation w/
students)
Varied data sources (observation data, parent feedback,
evaluator feedback, peer feedback, student work,
assessment results)
Information about individual needs of students (IPs,
etc.)
Logs of phone calls/parent contacts/emails
Lunch count
Field trip data
Assessment data
Student’s own data files (dot charts, learning progress,
graphs of progress, portfolios)
Interaction with PTA or parent groups or parent
volunteers
Daily assignment notebooks Requiring parents to
discuss and sign off on assignments
Proactive or creative planning for parent-teacher
conferences (including students in the process)
Asking parents how they think their student is doing
Regular parent “check-ins”
Take or teach school-level professional development
sessions
Collegial planning (not working in isolation)
Inviting people into your classroom
Using resources (specialists, support staff)
Treating people with respect
Supportive and collaborative relationships with
colleagues
2.
Optional
 Grade Book
 PD Plan
 Student / Parent Survey
 Observations
 Notes taken during observation
1.
Evaluator/Teacher conversations:
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
2.
3.
Optional
 Grade Book
 PD Plan
 Progress Reports
1. Logs of communication with parents
 Teacher log of communication (who, what,
why, when, “so what”?)
 Progress reports, etc.
1.
2.
3.
Observations
 Notes taken during observation
Attendance at PD sessions
Optional
 PLC agendas
 Evidence of community involvement
 Evidence of mentorship or seeking to be
mentored
44
4e: Growing and
developing
professionally








4f: Demonstrating
professionalism




Evaluator/Teacher
Conversations
Observation
Lesson/unit plan
Professional development plan
Mentoring involvement
Attendance or presentation at
professional organizations /
conferences / workshops / PLCs
Membership in professional
associations or organizations
Action research
-
Evaluator/Teacher
conversations
Observation of participation in
PLC meetings or school
leadership team meetings
Scheduling and allocation of
resources
School and out-of-school
volunteering
-
Academic inquiry, continuous improvement, lifelong
learning
1.
2.
3.
-
Obtaining additional resources to support students
individual needs above and beyond normal
expectations (i.e., staying late to meet with students)
Doing what’s best for kids rather than what is best for
adults
Mentors other teachers
Draws people up to a higher standard
Having the courage to press an opinion respectfully
Being inclusive with communicating concerns (open,
honest, transparent dialogue)
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
Lesson plans/unit plans
Observations
 Notes taken during observation
4.
Optional
 PD plan
 PLC agendas
 Evidence of participating in PD
 Evidence of mentorship or seeking to be
mentored
 Action research
1.
Evaluator/Teacher conversations
 Guiding questions
 Documentation of conversation (e.g.,
notes, written reflection, etc.)
2.
Optional
 Teacher provides documents to evaluator
at end of year/semester
 Written reflection
 Parent and student survey
 Observing teacher interacting with
peers/students/families
 Record of unethical behavior
45