Orwell Award Announcement SusanOhanian.Org Home


Outrages

 

9486 in the collection  

    Why NY's test mess is far from over

    Fred Smith, a retired test analyst and an independent whistle blower, reveals information about the New York State tests that nobody else talks about.

    By Fred Smith

    New York's "test mess" is worse than even the avowedly reformist state education leaders have acknowledged -- and it may not be over yet, either.

    A close look at the data (some of which became available only via the Freedom of Information Law) strongly suggests that the exams created each year by CTB/McGraw-Hill -- which purportedly measure the math and English proficiency of 1.2 million New York students -- are fundamentally flawed. That means that even Regents Chancellor Meryll Tisch and state Education Commissioner David Steiner's "recalibrating" of the scoring can't fix the problem.

    The state Education Department paid the company $38 million for the tests used in 2006-'09.

    Like most such tests, CTB's exams contain both multiple-choice and constructed-response questions. The latter ask students to produce a response, for example showing how they solved a math problem or writing answers to express their understanding of reading passages.

    Constructed-response items take more time and money to administer and score -- but educators generally believe these questions measure a higher order of knowledge and thought than multiple-choice items, which kids typically find less challenging.

    Yet results from both types of questions should point in the same direction -- that is, if this year's 4th graders do markedly better on the math multiple-choice questions than they did the year before, then they ought to improve on the math constructed-response items, too.

    In other words, on well-developed tests, the results on both types of questions are in harmony -- pointing in the same direction and nearly parallel from one year to the next. After all, each is supposed to tap a different level of knowledge of the same subject. Performance should move in a synchronized way.

    That's exactly the pattern shown on the National Assessment of Educational Progress -- nationally and in New York. The "nation's report card" uses both types of items to measure reading and math proficiency -- and the performance of New York kids on both is strikingly consistent over time.

    Not so, the results on the state exams.

    Consider just the math tests, administered every year to students in each of six grades. We have data on the four years from 2006 to 2009, so we can look at whether scores went up or down for the six grade levels in each of three school years -- 18 comparisons in total.

    In 10 of the 18 cases, raw scores (i.e., the percentage of questions answered correctly) rose on one of the types of question, but fell for the other. In four cases, there was a smaller divergence. In only four cases did the scores clearly move in the same direction.

    Any testing professional should recognize this as an alarm bell: Something is seriously wrong with these exams. (And it is the tests, not the students or anything else: Again, the NAEP exams, covering the same areas, do not show these bizarre divergences over time.)

    There are several more disturbing facts about the 2006-'09 exams:

    * Larger gains were usually made on multiple-choice items than constructed response. This boosted the overall score -- leading to press releases and headlines that suggested everything was improving.

    * Worse, data that contradict that storyline went undisclosed: The public didn't see separate analyses of constructed-response scores.

    * Statistics (obtained via the Freedom of Information Law) on the field tests (where questions get "tried out" prior to creating the actual exams) show inconsistencies between multiple-choice and constructed response items. CTB should have seen this data and realized it had a big problem.

    Internal consistency is a mark of test reliability -- and without reliability, tests can't measure anything in a valid way. And New York's exams have clearly been lacking in consistency.

    It's likely the just-released 2010 test results bear the same fatal flaw. The "solutions" on offer from Tisch and Steiner -- raising "cut scores" and increasing the scope of material on the exams -- don't address the overriding issue.

    What's needed is an independent probe of the testing program, one with sweeping authority to investigate the role of Education Department officials, CTB measurement specialists and the state's technical advisers in all aspects of the program.

    I believe we've been sold defective goods. For starters, we should demand our money back.

    Fred Smith, a retired Board of Education senior analyst, worked for the city public-school system in test research and development.


    — Fred Smith
    New York Post
    2010-09-08
    http://www.nypost.com/p/news/opinion/opedcolumnists/why_ny_test_mess_is_far_from_over_C3w9b48XKvPZq6cbZPTXeK


    INDEX OF OUTRAGES

Pages: 380   
[1] 2 3 4 5 6  Next >>    Last >>


FAIR USE NOTICE
This site contains copyrighted material the use of which has not always been specifically authorized by the copyright owner. We are making such material available in our efforts to advance understanding of education issues vital to a democracy. We believe this constitutes a 'fair use' of any such copyrighted material as provided for in section 107 of the US Copyright Law. In accordance with Title 17 U.S.C. Section 107, the material on this site is distributed without profit to those who have expressed a prior interest in receiving the included information for research and educational purposes. For more information click here. If you wish to use copyrighted material from this site for purposes of your own that go beyond 'fair use', you must obtain permission from the copyright owner.