Академический Документы
Профессиональный Документы
Культура Документы
Executive Director
Phone #
(916) 228-2612
swinlock@scoe.net
E-Mail
If
the
preparer
of
this
report
is
different
than
the
Program
Contact,
please
note
contact
information
for
that
person
below:
Name
Title
Program Manager
Phone #
916-228-2575
mneuburger@scoe.net
Although
the
program
offers
support
to
administrators
beyond
the
scope
of
the
credentialing
programs,
this
report
will
focus
only
on
analysis
of
the
credentialing
programs
and
the
common
and
program
standards
that
are
assessed.
2012-2013
Number
of
Number
of
Candidates
Completers/
Graduates
70
70
2013-14
Number
of
Number
of
Candidates
Completers/
Graduates
67
65
16
15
23
23
15
41
10
2
SECTION
A
CREDENTIAL
PROGRAM
SPECIFIC
INFORMATION
PART
II
Candidate
Assessment/Performance
and
Program
Effectiveness
Information
Candidate
Assessments
used
to
Recommend
Candidates
for
Credentials
a) Please
identify
and
describe
the
tool(s)
used
to
assess
candidates,
the
data
collection
process
and
the
types
of
data
collected
The
SCOE
Leadership
Institute
utilizes
several
primary
candidate
assessments
to
recommend
a
candidate
for
a
credential.
These
assessments
and
measures
focus
on
the
successful
completion
of
coursework
and
the
application
of
learning
modules
to
instructional
management
and
school
leadership.
Candidate
Assessments
Key
Assessment
Tools
Description
Standards
Assessed
March 2014
3
Candidate
Assessments
(continued)
Key
Assessment
Tools
Description
Standards
Assessed
CS
1,
2,
6,
7,
9
Additional
Data
and
Measures
Used
to
Analyze
Program
Effectiveness
and
Inform
Programmatic
Decision
Making
b) What
additional
information
about
program
effectiveness
is
collected
and
analyzed
that
informs
programmatic
decision-making?
In
addition
to
measuring
candidate
competencies,
the
SCOE
Leadership
Institute
also
assesses
additional
information
to
evaluate
and
inform
the
programs
effectiveness.
The
table
below
describes
the
variety
of
assessments
used
during
the
2012-13
and
2013-14
academic
years
that
informed
programmatic
decisions.
Data
was
collected
from
candidates,
coaches,
and
Leadership
Institute
staff
and
was
reviewed
throughout
each
year.
March 2014
4
Program
Effectiveness
Measures
of
Program
Effectiveness
Assessment
1.
Candidate
Evaluation
of
Course
Effectiveness
Assessment
2.
Workshop
Feedback
Assessment
3.
Reflective
Records
Assessment
4.
Mid-Program
Review
Description
Standards
Assessed
CS 2, 4, 9
CS 2, 3, 6
Candidate
Assessment
and
Program
Effectiveness
Data
c) Include
aggregated
data
from
4-6
assessments
that
were
described
in
(a)
and
(b).
Candidate
Assessment
1.
CPSEL
Coursework
Completion
Tier
I
Tier
I
candidates
are
required
to
complete
nine
specific
elements
of
the
program.
These
nine
elements
are:
Program
orientation
Completion
of
the
six
CPSEL
courses
and
related
coursework
Completion
of
the
end-of-year
project
and
symposium
plan
Presentation
of
the
end-of-year
project
at
a
symposium
March 2014
5
The
table
below
shows
the
candidate
counts
of
program
element
completion
by
location
and
year.
For
both
years,
program
completion
rates
were
very
high.
For
the
3
candidates
that
did
not
complete
the
program
within
the
expected
year:
Two
candidates
dropped
out
of
the
program
after
completing
only
the
orientation
in
2013-14.
One
candidate
in
2012-13
had
to
postpone
the
completion
of
the
project
and
symposium
due
to
severe
family
circumstances.
The
candidate
is
scheduled
to
complete
the
program
in
the
2014-15
year.
2012-2013
2013-2014
7
of
9
9
of
9
1
of
9
9
of
9
Elements
Elements
Elements
Elements
70
2
65
1
15
23
Grand
Total
1
85
2
88
98.8%
97.8%
2012-13
Cohort
(N=70)
2013-14
Cohort
(N=67)
Two
Two
Three
Three
Grand
Total
10
March 2014
17
6
Total
12
18
38
Total
12
19
22
50
Grand Total
Tier
II
The
table
below
shows
the
number
of
courses
completed
by
Tier
II
candidates
as
of
August
2014.
Although
the
candidates
can
enter
the
program
at
any
of
three
points
during
the
year
(fall,
winter,
or
spring),
each
candidate
is
expected
to
complete
three
CPSEL
courses
each
year
to
stay
on
track.
Of
the
2012-13
candidates
that
completed
zero
courses,
one
dropped
out
and
the
other
had
to
put
the
program
on
hold
due
to
medical
reasons.
Of
the
2013-14
candidates
that
completed
zero
courses,
all
four
had
begun
the
first
CPSEL
course
in
May
of
2014.
Overall
since
the
beginning
of
the
program,
10
candidates
from
the
2012-13
cohort
have
completed
the
program
and
received
their
clear
credential.
Tier
II
Candidates:
Number
of
CPSEL
Courses
Completed
as
of
August
2014
0
courses
1
courses
2
courses
3
courses
4
courses
5
courses
6
courses
Grand
Total
2013-14
4
13
9
10
36
Candidate
Assessment
2.
End-of-Year
Project
and
Symposium
The
end-of-year
project
and
symposium
is
an
evolving
process
whereby
the
district
coaches
and
faculty
mentors
work
with
the
Tier
I
candidates
to
provide
continuous
feedback
on
this
element
of
the
program.
This
occurs
through
one-on-one
meetings
and
assessment
of
the
progress
made
along
the
way.
At
the
end
of
the
year,
the
candidates
present
their
project
to
a
panel
of
educational
leaders
from
the
region
and
receive
feedback
and
evaluation
on
their
projects.
Panel
members
in
each
of
the
following
required
elements
assess
projects
and
presentations:
Project
vision
and
expected
outcomes
Rationale
of
need
and
importance
Evidence
of
transformational
leadership
Implementation
artifacts
Progress
towards
accomplishing
the
vision
Next
steps
March 2014
7
The
feedback
of
these
elements
(captured
through
a
locally
developed
form)
is
provided
to
candidates.
Of
all
the
candidates
that
remained
in
the
program
in
2012-13
and
2013-14,
only
one
did
not
complete
the
end-of-year
project
and
symposium
as
noted
above
and
is
scheduled
to
complete
it
in
2014-15.
A
sample
of
comments
from
two
of
the
candidates
indicated
that
the
program
provided
a
chance
to
engage
in
leadership
roles
and
to
act
as
a
mentor
to
new
teachers.
It
also
afforded
the
opportunity
to
learn
how
to
assess
and
reassess
successes
and
to
value
data
and
use
feedback
to
foster
improvement.
Candidate
Assessment
3.
Descriptions
of
Practice
Assessments
Throughout
the
duration
of
the
program,
Tier
II
candidates
utilize
the
Moving
Leadership
Standards
into
Everyday
Work
Descriptions
of
Practice
tool
to
assess
their
level
of
competency
in
each
of
the
six
program
standards
measured.
Candidates
complete
this
assessment
initially
at
the
beginning
of
the
program,
again
after
completing
one
year
of
the
program,
and
lastly
at
the
end
of
the
program.
All
participants
are
required
to
complete
these
assessments
therefore
the
participation
rate
was
100%.
Scoring
rubric
used
for
analyses:
1
=
directed
toward
the
standard
2
=
approaches
the
standard
3
=
meets
the
standard
4
=
exemplifies
the
standard
In
addition
to
assessing
their
level
of
competency,
candidates
provided
comments
on
their
growth
experience
and
described
the
practices
they
employed.
The
table
below
shows
the
average
scores
for
each
area
at
each
stage
of
measurement
as
well
as
the
number
of
candidates
that
grew
one
or
more
levels
at
the
mid-point
and
end-point.
Initial
Program
Mid-
Program
End
of
Program
Number
Increased
1
or
More
Levels
Mean
(N
=
50)
Std.
Dev.
Mean
(N
=
17)
Std.
Dev.
Mean
(N
=
10)
Std.
Dev.
Initial
to
Mid
Initial
to
End
2.08
2.18
.900
.825
2.71
2.59
.588
.507
3.00
3.10
.667
.568
11
11
7
7
1.90
.814
2.76
.562
3.10
.568
13
2.18
2.24
2.20
.850
.822
.728
2.82
2.88
3.06
.529
.332
.556
3.20
3.20
3.20
.632
.632
.632
10
13
13
6
7
8
1.98
.769
2.47
.514
3.10
.568
10
March 2014
8
3.1
Ensure
a
safe
school
environment
2.38
.805
2.88
.600
3.40
.699
2.14
.833
2.65
.493
3.20
.632
2.24
2.34
.771
.982
2.76
3.12
.437
.332
3.20
3.30
.422
.483
8
11
6
6
2.10
.814
2.71
.772
3.30
.675
11
1.84
.766
2.47
.800
2.80
.632
1.86
.700
2.41
.870
2.70
.675
2.76
.960
3.06
.429
3.60
.516
2.18
.850
2.71
.470
3.20
.632
11
2.66
.823
3.00
.500
3.40
.516
2.50
.909
3.12
.600
3.50
.527
1.88
.824
2.29
.772
2.70
.483
1.98
1.88
.742
.718
2.82
2.29
.809
.686
2.80
2.80
.422
.632
13
5
6
6
Candidate
Assessment
4.
Exit
Interview
After
Tier
II
candidates
complete
each
of
the
six
CPSEL
courses
and
additional
program
requirements,
they
must
complete
the
exit
interview
to
clear
their
administrative
services
credential.
The
exit
interview
included
a
practicum
review,
a
description
of
practice
review,
and
a
coaching
review.
Feedback
from
candidates
during
the
interviews
was
collected
and
was
used
to
assist
candidates
in
reflecting
on
their
program
experience
and
personal
and
professional
growth.
Of
the
10
candidates
that
began
the
program
in
fall
of
2012,
100%
had
completed
the
exit
interview
and
had
completed
the
program.
An
analysis
of
each
review
area
indicated
that:
100%
of
candidates
met
the
standards
in
all
6
CPSELs
100%
of
candidates
were
rated
by
the
coaches
as
exceeding
the
requirements
Overall
coaches
indicated
that
candidates
demonstrated
great
growth
over
the
course
of
the
program,
often
moving
from
approaching
standards
to
meeting
or
exceeding
them.
Areas
noted
for
continued
growth
for
candidates
included
expanding
relationships
with
parents
and
the
community
and
also
refining
their
leadership
and
management
skills.
March 2014
9
Program
Effectiveness
Assessment
1.
Candidate
Evaluation
of
Course
Effectiveness
After
the
completion
of
each
CPSEL
course,
Tier
I
candidates
were
asked
to
rate
the
course
effectiveness
by
indicating
their
level
of
agreement
for
each
of
10
statements.
Scoring
rubric
used
for
analyses:
1
=
strongly
disagree
(was
coded
as
2
on
original
survey)
2
=
disagree
(was
coded
as
3
on
original
survey)
3
=
agree
(was
coded
as
4
on
original
survey)
4
=
strongly
agree
(was
coded
as
5
on
original
survey)
The
participation/completion
rate
for
the
workshop
feedback
forms
for
all
Tier
I
candidate
groups
and
locations
was
100%
for
both
2012-13
(N=86)
and
2013-14
(N=87).
The
table
below
shows
the
results
by
year
and
location
with
the
responses
for
agree
and
strongly
agree
combined
to
demonstrate
the
percent
that
agree.
Cells
highlighted
in
yellow
indicate
where
less
than
93%
of
candidates
agreed
with
the
statements.
Green
shaded
cells
indicate
areas
where
100%
of
candidates
agreed
with
the
statements.
When
computed
as
averages
on
the
scale
from
1
to
4,
the
means
for
each
course
by
location
and
year
ranged
from
3.46
to
3.87
and
the
standard
deviations
ranged
from
.485
to
.712.
March 2014
10
Mgmt.
for
Teaching
Personal
Ethics
Political
Influences
Shared
Vision
Learning
Culture
Diverse
Families
and
Comm.
Year
% Agree
% Agree
% Agree
% Agree
% Agree
% Agree
2012-13
2013-14
97.1%
98.4%
97.1%
98.4%
97.1%
98.4%
97.1%
98.4%
97.1%
98.4%
100.0%
98.4%
100.0%
98.4%
100.0%
98.4%
94.3%
95.2%
94.3%
98.4%
98.5%
100.0%
100.0%
100.0%
100.0%
100.0%
95.5%
98.4%
97.0%
100.0%
97.1%
96.9%
97.1%
96.9%
97.1%
96.9%
97.1%
96.9%
97.1%
96.9%
100.0%
100.0%
100.0%
100.0%
100.0%
98.4%
100.0%
100.0%
100.0%
100.0%
97.1%
96.8%
97.1%
96.8%
97.1%
96.8%
97.1%
96.8%
97.1%
96.8%
2012-13
97.1%
98.6%
97.0%
97.1%
100.0%
97.1%
2013-14
98.4%
98.4%
100.0%
96.9%
100.0%
96.8%
2012-13
2013-14
97.1%
98.4%
97.1%
98.4%
94.3%
96.7%
95.7%
93.7%
97.0%
98.4%
97.0%
100.0%
97.1%
96.9%
97.1%
96.9%
100.0%
98.4%
98.5%
100.0%
97.1%
96.8%
98.6%
96.8%
2012-13
87.1%
97.1%
91.0%
98.6%
94.2%
94.3%
2013-14
98.4%
98.4%
100.0%
90.6%
96.8%
93.7%
92.9%
92.9%
92.5%
98.6%
92.6%
94.3%
96.9%
98.4%
100.0%
95.3%
93.7%
95.2%
Shasta COE
Year
% Agree
% Agree
% Agree
% Agree
% Agree
% Agree
2012-13
100.0%
95.0%
93.8%
95.0%
100.0%
95.0%
93.8%
95.0%
93.8%
95.0%
86.7%
100.0%
86.7%
95.0%
86.7%
100.0%
86.7%
100.0%
86.7%
100.0%
100.0%
86.4%
100.0%
86.4%
100.0%
86.4%
100.0%
86.4%
100.0%
86.4%
100.0%
100.0%
100.0%
100.0%
100.0%
100.0%
100.0%
100.0%
100.0%
100.0%
93.8%
100.0%
93.8%
100.0%
93.8%
100.0%
93.8%
95.2%
93.8%
90.5%
2013-14
2012-13
2013-14
2012-13
2013-14
2012-13
2013-14
2012-13
2013-14
2012-13
2012-13
2013-14
100.0%
95.5%
100.0%
95.5%
100.0%
95.5%
100.0%
95.5%
100.0%
95.5%
2012-13
100.0%
93.8%
86.7%
100.0%
100.0%
93.8%
2013-14
95.5%
95.0%
100.0%
86.4%
100.0%
95.2%
2012-13
2013-14
100.0%
95.5%
100.0%
95.5%
93.8%
95.0%
100.0%
95.0%
86.7%
100.0%
86.7%
100.0%
100.0%
86.4%
100.0%
86.4%
100.0%
100.0%
100.0%
100.0%
93.8%
100.0%
93.8%
100.0%
2012-13
93.8%
100.0%
93.3%
100.0%
100.0%
93.8%
2013-14
95.5%
95.0%
100.0%
86.4%
100.0%
100.0%
2012-13
10
93.8%
93.8%
86.7%
100.0%
100.0%
87.5%
2013-14
95.5%
95.0%
100.0%
86.4%
100.0%
95.2%
2013-14
2012-13
2013-14
2012-13
2013-14
2012-13
2013-14
2013-14
2012-13
11
Program
Effectiveness
Assessment
2.
Workshop
Feedback
Tier
II
candidates
were
required
to
complete
five
workshops
of
their
choice
in
each
year
of
the
program.
At
the
conclusion
of
each
workshop
they
completed
a
workshop
feedback
form
that
assessed
the
quality
of
the
presenter
and
the
workshop.
Participants
were
asked
to
indicate
their
level
of
agreement
to
seven
statements
and
provide
comments
about
the
workshop
and
the
application
of
the
workshop
to
their
own
work.
The
table
below
shows
the
average
score
and
descriptive
statistics
for
each
statement
by
workshop
year
and
includes
information
about
workshop
length
and
diversity.
Scoring
rubric
used
for
analyses:
1
=
strongly
disagree
2
=
disagree
3
=
neutral
4=
agree
5
=
strongly
agree
Workshop
comments
indicated
that:
The
information
was
valuable
and
useful
Presenters
were
very
knowledgeable
Opportunities
to
collaborate
with
colleagues
were
appreciated
Comments
regarding
the
applicability
of
workshop
content
to
participants
work
generally
indicated
that
seeing
how
the
presenter
applied
the
information
in
various
settings
helped
them
to
understand
the
challenges
in
applying
it
to
their
work.
2012-13
Tier
II
Workshop
Evaluations
by
Candidates
2013-14
Mean
Std.
Dev.
4.58
60
4.63
Instructor Effectiveness
Workshop Content
Mean
Std.
Dev.
.591
4.78
189
.509
60
.551
4.75
189
.555
4.67
60
.572
4.77
189
.545
4.55
60
.675
4.70
189
.564
4.50
60
.676
4.75
189
.555
4.27
60
.778
4.44
189
.724
4.50
60
.701
4.57
189
.708
Additional
Information
Average
Workshop
Length
Number
of
Different
Workshops
March 2014
11
3.94 hours
4.74 hours
39
116
12
Program
Effectiveness
Assessment
3.
Reflective
Records
Throughout
the
Tier
II
program,
coaches
and
candidates
reflect
on
the
program
support
process
at
the
conclusion
of
each
CPSEL
course.
The
amount
of
program
support
is
rated
across
6
areas.
The
table
below
shows
the
average
rating
for
each
CPSEL
course
by
year
and
by
candidates
and
coaches.
Scoring
rubric
used
for
analyses:
1
=
none
2
=
minimal
3
=
adequate
4=
over
and
above
Program
Year
Tier
II
CPSEL
Courses:
Overall
Average
Score
for
Support
Given
or
Received
Candidate
Reflective
Record
Shared
Vision
of
Learning
Teaching
and
Learning
Culture
Management
for
Teaching
and
Learning
Working
with
Diverse
Families
and
Communities
Personal
Ethics
for
Leadership
Political
and
Social
Influences
Coach
Reflective
Record
Shared
Vision
of
Learning
Teaching
and
Learning
Culture
Management
for
Teaching
and
Learning
Working
with
Diverse
Families
and
Communities
Personal
Ethics
for
Leadership
Political
and
Social
Influences
2012-13
2013-14
Mean
Count
Std.
Dev.
Mean
Count
Std.
Dev.
3.63
3.48
3.31
10
11
9
.47
.85
.94
3.83
3.28
3.47
3.43
4
23
35
46
.33
.68
.54
.62
3.55
3.39
3.39
10
14
12
.48
.52
.51
3.08
3.24
3.09
3.11
2
25
35
46
.12
.49
.47
.53
To
assess
the
strength
of
the
support
received,
each
item
on
the
reflection
tool
was
reviewed
by
year.
The
charts
that
follow
depict
these
results
for
candidates
and
coaches
across
all
CPSEL
courses
combined.
Also
note
that
coaches
are
expected
to
visit
the
candidate
at
their
site
at
least
one
time
per
year
(this
equates
to
one
visit
during
one
of
the
three
CPSEL
courses
annually,
hence
a
33%
visitation
rate).
March 2014
12
13
March 2014
13
14
Program
Effectiveness
Assessment
4.
Mid-Program
Review
After
completing
one
year
of
the
program,
Tier
II
candidates
complete
a
locally
developed
mid-
program
form
that
requests
feedback
about
the
program
quality
and
their
experience
as
a
candidate.
This
form
is
designed
to
capture
free-form
comments
in
four
key
areas
using
the
following
guiding
questions:
Program
o How
are
practicums
aligned
with
elements
of
standards?
o What
knowledge
areas
will
you
focus
on
for
remaining
CPSELS?
Coaching
o What
are
some
examples
of
coaching
strategies
that
have
supported
your
leadership
development?
Logistics
o What
contacts
are
being
made
that
support
development?
o What
areas
of
coaching
would
be
helpful
to
further
develop
leadership?
Support
o What
assistance
has
been
provided
during
the
program?
o What
area(s)
of
need
would
assist
the
work
needed
to
complete
the
program?
A
total
of
10
candidates
from
the
2012-13
cohort
and
14
from
the
2013-14
cohort
completed
this
requirement
to
date.
The
responses
provided
through
these
forms
were
reviewed
by
Leadership
Institute
staff
and
are
locally
stored.
Because
a
lengthy
analysis
of
all
comments
received
would
not
serve
the
concise
nature
of
this
report,
a
review
of
a
representative
sample
from
the
2013-14
is
provided
instead.
Overall,
data
collected
from
the
mid-program
review
indicated
that:
Program
Candidates
emphasized
how
practical
the
program
work
was
and
how
relevant
it
was
to
their
work.
Goals
for
the
school
site
they
worked
at
meshed
well
with
the
expectations
of
the
leadership
program.
The
practicums
were
described
as
authentic
and
beneficial
to
candidates
learning
and
professional
growth.
Coaching
Candidates
were
very
thankful
for
the
support
and
feedback
they
received
from
their
coaches
and
the
validation
of
their
work
in
the
program.
Coaching
sessions
were
timely
and
relevant
and
the
discussions
helped
candidates
explore
options
and
expand
their
perspectives.
Logistics
Candidates
agreed
that
the
program
logistics
were
flexible
and
that
the
schedules
provided
them
with
sufficient
time
to
complete
their
tasks.
Support
Overall
districts
were
very
supportive
of
their
candidates
including
the
provision
of
financial
support
for
the
program.
Sites
appreciated
the
professional
development
activities
that
candidates
brought
back
to
their
site
and
the
Leadership
Institute
was
noted
for
its
overall
organization
of
the
program
and
the
availability
and
responsiveness
of
staff.
March 2014
14
15
PART
III
Analyses
and
Discussion
of
Candidate
and
Program
Data
Describe
what
the
analyses
of
the
data
demonstrate
about
your
program
relative
to:
a)
candidate
competence;
and
b)
program
effectiveness.
Overall,
results
from
the
candidate
and
program
assessments
indicate
that
the
Leadership
institute
provides
appropriate
professional
development
and
support
to
administrative
credential
candidates
to
enable
them
to
successfully
complete
the
program
requirements
and
earn
the
related
credentials.
15
16
rated
themselves
at
the
same
level
or
one
level
lower
at
the
conclusion
of
the
program,
likely
due
to
a
better
knowledge
of
what
exemplifying
a
standard
truly
means.
This
could
be
considered
a
strength
in
the
program
whereby
candidates
are
gaining
a
better
understanding
of
the
leadership
standards
and
expectations
as
they
progress
through
the
program.
This
also
attests
to
candidates
ability
to
effectively
reassess
their
learning.
Candidate
Assessment
4.
Exit
Interview
Data
from
the
Tier
II
exit
interview
indicated
that
the
candidates
all
met
each
standard
and
made
great
growth
however
the
coaches
comments
regarding
areas
of
significant
growth
and
areas
for
continuous
growth
were
general
in
nature.
More
specific
comments
with
concrete
examples
as
well
as
resources
and
suggestions
for
improvement
would
benefit
the
professional
growth
of
candidates
and
review
by
program
staff.
It
is
also
worth
noting
that
each
candidate
was
rated
as
exceeding
the
standards
of
the
program
which
supports
strong
candidate
competency.
On
the
other
hand,
this
might
be
erroneously
interpreted
by
the
candidates
that
they
have
reached
their
full
potential
as
an
administrator
and
that
little
growth
is
left
to
accomplish.
However,
given
the
comments
by
candidates
of
their
personal
and
professional
growth,
this
is
likely
not
a
concern.
Nonetheless,
a
rating
scale
with
more
levels
could
be
developed
to
provide
this
assurance
and
refine
this
assessment
of
candidate
competency.
Program
Assessment
1.
Candidate
Evaluation
of
Course
Effectiveness
Overall,
the
Tier
I
courses
received
high
ratings
for
overall
effectiveness.
The
highest
rated
areas
were
the
Learning
Culture
and
the
Political
and
Social
Influences
across
both
program
sites
and
for
both
years.
Political
and
Social
Influences
had
improved
at
the
Shasta
COE
site
from
the
prior
year
and
the
online
learning
component
of
this
course
improved
at
both
sites
from
2012-13
to
2013-14.
Beyond
those
areas,
the
program
sites
differed.
For
example,
the
Management
for
Teaching
course
was
rated
higher
at
the
Shasta
COE
site
than
at
the
SCOE
site
while
the
rating
for
the
Shared
Vision
course
at
the
Shasta
COE
had
dropped
from
2012-13
to
2013-14.
Because
different
staff
teach
these
courses
at
the
different
sites,
it
would
be
valuable
to
learn
what
is
different
about
how
these
courses
are
being
taught.
Nonetheless,
the
ratings
for
all
courses
were
very
high
and
support
the
strength
of
the
program.
Program
Assessment
2.
Workshop
Feedback
At
the
conclusion
of
each
workshop,
Tier
II
candidates
completed
workshop
evaluation
forms.
As
evidenced
by
the
data,
information
from
the
2012-13
evaluations
was
successfully
used
to
improve
all
workshop
areas
for
the
2013-14
year.
This
improvement
also
coincided
with
the
increased
lengths
of
the
workshops
by
nearly
an
hour,
addressing
the
desire
by
candidates
to
go
more
in-depth
during
courses.
The
areas
of
greatest
strength
in
the
2013-14
year
for
the
workshops
was
in
the
instructors
demonstrated
knowledge
of
key
instructional
components
and
usefulness
of
materials
in
meeting
the
objectives
of
the
course.
The
area
of
relative
weakness
for
both
years
was
in
the
candidates
belief
that
they
were
prepared
to
implement
the
workshop
content,
a
common
theme
found
in
many
professional
development
programs.
March 2014
16
17
Program
Assessment
3.
Reflective
Records
Results
for
both
program
years
were
similar
and
it
was
interesting
to
note
that
Tier
II
candidates
reported
receiving
slightly
more
support
from
the
program
coaches
than
the
coaches
believe
they
provided.
This
was
most
evident
in
the
Management
for
Teaching
and
Learning
course
which
received
the
highest
rating
during
the
2013-14
year
yet
received
the
lowest
rating
in
the
2012-13
year.
Results
also
indicated
that
candidates
need
greater
coach
support
in
the
Working
with
Diverse
Families
and
Communities
course.
A
comparison
of
candidate
ratings
to
coach
ratings
across
all
CPSEL
courses
shows
that
in
2013-
14
a
disparity
between
perceived
support
offered
and
received
existed
across
nearly
every
area
measured.
In
2013-14
communication
through
a
variety
of
mediums
was
the
highest
rated
area
of
support
by
candidates
where
68%
of
candidates
indicated
that
coaches
went
over
and
above
in
this
area
with
the
next
highest
support
areas
identified
as
assisting
with
and
reviewing
the
Applied
Practicum
Action
Plan.
Although
the
area
of
lowest
support
would
appear
to
be
in
the
area
of
visiting
the
site,
this
may
not
be
the
case.
It
is
difficult
to
discern
from
the
structure
of
the
question
whether
it
is
assessing
whether
a
site
visit
occurred,
whether
several
site
visits
occurred,
or
whether
a
visit
occurred
and
the
coach
was
particularly
supportive
during
the
visit.
The
question
stem
was
whether
the
coach
supported
the
candidate
by
visiting
the
site
however
this
is
based
on
a
four
point
rubric
ranging
from
none
to
above
and
beyond.
Interpretation
of
the
data
gleaned
from
this
question
is
inconclusive.
It
is
recommended
that
this
item
be
scored
on
a
yes/no
rubric
in
the
future.
On
an
interesting
note,
it
should
have
occurred
that
coaches
became
more
experienced
with
the
program
over
time
and
the
coaches
in
the
second
year
and
should
have
been
able
to
provide
better
support
to
candidates,
yet
coaches
indicated
that
they
gave
less
support
in
the
second
year
of
the
program.
Again,
it
is
difficult
to
interpret
whether
coaches
actually
provided
less
support
or
whether
they
became
more
efficient
in
their
delivery
of
it.
Program
Assessment
4.
Mid-Program
Review
Analysis
of
the
mid-program
review
data
confirms
the
findings
of
the
other
program
assessments.
Overall
the
program
is
efficiently
run
and
the
content
and
format
are
relevant
to
the
work
of
the
candidates.
Based
on
the
overall
information
provided
by
candidates,
no
causes
for
program
improvement
were
discovered
however
a
revised
review
tool
that
includes
a
rating
scale
might
be
able
to
better
discern
subtle
areas
where
the
program
could
benefit
from
refinement.
17
18
the
program,
most
of
which
are
a
refinement
of
assessment
tools
used
and
the
increased
sharing
of
the
assessment
results
to
candidates
and
other
stakeholders.
The
table
below
presents
the
focus
areas
for
improvement
and
the
proposed
changes
to
support
improvements.
Data
Source
Plan
of
Action
or
Proposed
Changes
Made
Applicable
Program
or
Common
Standard(s)
Candidate
To
better
assess
the
quality
of
the
coursework
that
Tier
I
CS
1,
2
Assessment
1.
candidates
complete,
the
scoring
rubric
will
be
changed
from
Meets/Does
not
Meet
Standards
to
a
multi-level
CPSEL
rubric
reflective
of
the
Description
of
Practice.
This
will
Coursework
provide
more
information
to
the
program
staff
about
Completion
candidate
competency
as
well
as
differences
seen
among
for
Tier
I
courses
or
instructors.
The
revised
rubric
will
be
pilot
tested
in
the
2014-15
year
and
implemented
in
the
2015-16
year.
Candidate
Details
about
the
project
and
symposium
and
the
exit
CS
1,2,6,7,9
Assessments
2
interview
will
be
captured
electronically
and
analyzed
and
4:
End-of- annually.
In
addition,
rating
scales
will
be
developed
and
candidates,
faculty,
and
Institute
staff
will
rate
their
Year
Project
effectiveness
in
meeting
course
elements
and
program
and
effectiveness.
This
information
will
be
used
to
guide
Symposium
program
design,
services
and
support.
and
Exit
Interview
Program
Assessment
2.
Workshop
Feedback
18
CS 6,7
CS 6,9
19
Program
Assessment
4.
Mid-Program
Review
Analysis
of
all
Candidate
and
Program
Assessments
All
Candidate
and
Program
Assessments
March 2014
19
CS 9