Recently I was looking at the NSRE math scores of some schools around the
state, and it occurred to me to share with you the results from a few
schools I've seen who's NSRE trends exemplify what we all hope to achieve.
At the same time I thought I would share some miscellaneous thoughts about
factors I consider when interpreting scores.

Jim Abrams

--------------------------------------------------------------------


1. The DOE School Improvement Website  When you are in the process of
looking at your NSRE math results and considering action planning, I can't
say enough good things about the DOE's School Improvement Support 'website'.
If you haven't been there you owe it a visit.
http://data.ed.state.vt.us/apg/index.html .  Not only can you look at your
own scores a number of ways, but you can also look at comparisons with some
other schools.

2. Reading Scores: Examples of score trends you would like to see. I am
going to say a few words about reading scores that you may already know, or
maybe not.  In doing so I want to use a few schools as examples, schools
who's scores look very promising at first inspection.  To know for sure
whether they clearly represent improved teaching, I would have to talk with
the schools more.  But they certainly represent the take we all would like
to see. Below are listed several elementary schools, a middle school and a
high school. Look at them on the DOE site  in conjunction with the narrative
below ( Use the question: Are we making progress toward all students meeting
or exceeding state standards?).  These schools are not the only schools with
impressive trends.  They simply happen to be some good examples I noticed
recently.

Lowell Village (4)
Derby Elementary (4)
Manchester Elementary (4)
Floodbrook (8)
Northfield High (10)

Measurement Error. When you look at scores there are several things you want
to consider.  First is 'measurement error'.  On the DOE site its shown as a
confidence interval. [a little bar above and below the actual score with a
line in between].  Schools with a small N will have a large confidence
interval.  The 'true score' for your school could be anywhere in that
interval.  With large confidence intervals what you need to look at is the
scores over time.  That will tell you in general whether you are looking at
reasonably 'true score' in a given year or a one year variation due to
random measurement factors.  You can't ever tell for sure in any year, but
looking at a trend in scores representing small N's like Lowell's (above)
you gain some confidence that their performance is not random score
variations.

If you have a school with a very small N (under 20) and patterns don't
appear as cleanly as one of the sample schools above, I would do a rolling
average.  Take the average of the first year and the second, and compare
that to an average of the second year and the third, the third and the
fourth etc.  The trend line produced by this technique reduces the impact of
measurement error but at the same time it also dampens the relative changes.
Still, if there was too much noise with small N's to determine much of
anything, I would use this technique and then look for a positive slope in
the overall trend line.

Cohort Effect Subjectively, everyone who has taught has experienced the
grade group ("class") of kids that was substantially better or worse
academic performers than usual.  This is what is called a cohort effect.  I
heard a math specialist at a department of education in a nearby state say
he believed that classes with up to a hundred students could exhibit
noticeable cohort effect on a state test. I am keeping an open mind about
his conjecture. In my own experience I have clearly seen a significant
cohort effect in a class of 60 students.

I have two ways I look for cohort effect. First, are scores in a year
unusual high or low compared to other years.  Are all three scores for that
year comparably high (or low). When you see an unexpected jump you always
have to ask two questions.  What do the teachers say about 'the class'.  Was
it an especially good or poor performing class. [by the way, this
qualitative assessment is best to get before teachers see the scores in
question]  The second thing to ask is "was there anything significant that
changed in the school in the year in question or the preceding year.

I have seen a high school exhibit a 23 point gain in a single year in a
class with 60 students.  The teachers had predicted they would see a bump
that year.  And they predicted that the gain would not be sustained the next
year.  Both prediction came true.  It was clearly a 'cohort effect'.

In a different case I saw a 20 point gain in students achieving or exceeding
the standards in concepts in a single year that represented a different kind
of change. It was in a school with an  N above 100.  Looking closely at
changes in the school the most likely explanation we found was a single
outstanding eighth grade teacher.  I wouldn't have guessed it was possible
but that explanation fit the data well, and much better than any alternative
hypothesis that we could produce.

Socio-economic effect. Another factor that can effect scores is SES,
basically poverty and affluence.  The single largest predictor of academic
performance is the education level of a student's mother.  This is in itself
correlated with income level.  Good scores from a school don't necessarily
represent good teaching compared to another school.  It may say much more
about the students.

If you want to look at a strong indicator of teaching power, look at
progress since the new standards exam was initiated.  If you can see
movements from 15-30 percent of your students making or exceeding standards
in concepts or problem solving up into the realm of 50-70 percent, with a
trend that looks trustworthy, then you are most likely looking at some
teaching power.  Simply having some of the highest scores in the state that
are relatively flat over time may reflect the capability of the students
much more than the teaching.

Changes in the Test: A last factor that can effect scores is changes in the
test form used by New Standards.  Its unfortunate that that happens, but
with the exception of the grade 10 scores a couple of years ago, I believe
the effect is relatively small, less than say 5-7 percentage points.  The
assessment folks at the DOE can speak much more knowledgeably about this.

The same test form was used in 1999 and 2000. Sometimes I use those years to
check whether a relatively modest change in scores in a school with a large
N may be attributable to the test rather than 'real' improvements in
performance.

Trends.  In the end, if you look at much of what has been said, it is the
trend data that tells the most about a schools scores and obviously
improvements in scores.  We are fortunate now to have a sufficient number of
years to see progress even in schools with small class sizes.  And we are
fortunate to have the DOE School Improvement Support Site.   Look at the
schools I have listed above.  They are examples of NSRE trends you would
like to see.  These schools suggest what is possible, not only in the clear
progress trends they demonstrate but also the absolute levels they have
reached to date.

They give us hope.  And they challenge us to further efforts.


'Jim Abrams

*********************************************
Jim Abrams
Director of Mathematics Education
Vermont Institute for Science, Math and Technology
7 West Street
Montpelier, Vt   05602
Ph: (802) 828-0069
[log in to unmask]