|
Module Seven -- Interpretation & Communication of Assessment Results |
|
|
|
|
|
|
Methods of Interpreting Test Scores Types of Test Scores Defined Purpose Interpretation of Cumulative Percentages Cautions in Interpreting Any Test Score Interpreting Scale Scores
The scores from achievement, aptitude, attitude, psychological tests can not be compared directly with each other unless the norming group is taken into consideration and the scale on which the score is based. We addressed this issue in earlier modules on aptitude and achievement tests. Methods
of Interpreting Test Scores
Are there enough items for each skill tested? What is the difficulty level of the items? What type(s) of items are used? What is the match of items to objectives? 2. Norm-Referenced Interpretation
Are the test norms representative? Are the test norms up to date? Are the test norms comparable? Are the test norms adequately described? Types of Test Scores and Defined Purpose
Grade Equivalent scores -- grade group in which student's raw score is average; used to estimate or monitor growth Standard scores -- terms of standard distance of student's raw score from the mean (average) in terms of standard deviations; used to monitor growth; better at reflecting reality than grade equivalent scores Normal Curve Equivalent -- a normalized standard score; used to avoid problems with grade equivalent scores and used to describe group performance and to show growth over time Percentile Ranks -- student's relative position in a group in terms of the percentage of students scoring lower than or equal to that student; used to determine relative areas of strengths and weaknesses; can create profile analyses from these scores. National Stanines -- normal distribution is divided into nine parts; used to identify relative areas of strengths and weaknesses Scale Scores -- scores on an arbitrarily
set common scale; used to measure students' progress across grades in a
subject
Interpretation of Cumulative Percentages
Standard scores are called standard because they have a constant mean and a constant standard deviation. According to this definition, which of the seven scores above are standard scores? Cautions in Interpreting Any Test Score
2. A test score should be interpreted in light of all of the student's relevant characteristics. 3. A test score should be interpreted according to the type of decision to be made. 4. A test score should be interpreted as a band of scores rather than as a specific score. 5. A test score should be verified by supplementary evidence. 6. Do NOT interpret a grade equivalent score as an estimate of the grade where a student should be placed. 7. Do NOT assume that the units are equal at different parts of the scale. 8. Do NOT assume that scores on different tests are comparable. 9. Do NOT interpret extreme scores as dependable estimates of a student's performance.
Look at the figures 17.3 and 17.4 in your text (pages 461 and 463). Notice that the score is presented as a band rather than a point. This is much more accurate and communicates clearly the approximate amount of error the publishers have found in their test. This is an estimate of the standard error of measurement. The "true" score, the students true ability in this area, will fall somewhere within the band. It may be near the top of the band, it may be in the middle of the band, or it may be near the end of the band. We do not know exactly where the true score lies, but we assume that the true score for that person lies somewhere within the band. Remember, people create tests, people are not perfect, tests are not perfect. All tests contain some error. The length of the band reflects this error. The longer the band the more error. Remember from the module on reliability that error reduces reliability. Therefore, the longer the band, the more error, and the less reliable the test is. The band is also helpful in determining true differences between scores. Remember, that we do not know exactly where the "true" score lies for this student, but we know it lies somewhere within the band. Using the student profile on page 461 of your text, determine the range of scores for this student in the areas of numerical reasoning and perceptual speed & accuracy. Do you see that the scores for both could be exactly the same depending on where the students "true" score is? Scores can not be the same for numerical reasoning and abstract reasoning. So there are real differences between the students scores on numerical reasoning and on abstract reasoning. In other words, when the bands overlap common area along the continuum, the scores can not be interpreted as being really different from each other. It is only when the bands do not share common area, that the scores can be interpreted as being really different from each other.
Is this student's score for verbal reasoning: 1. above average
Is this student's score for abstract reasoning: 1. above average
Given a scale score and the number of students earning below that and the number of students earning exactly that scale score the cumulative frequency at midpoint can be calculated. The definition of cumulative frequency at midpoint is all the students who earned scale scores below a given score plus one half of the students who earned that scale score. An example, if we know that 36 students earned scale scores lower than 400 and 6 students earned scale scores of exactly 400, then we take one half of 6 and add that to 36, and we know that the cumulative frequency at midpoint for a scale score of 400 is 40. If we then divide that by the number of students who took the test, we have the percentile rank. Given that 50 students took the test, we divide 40 by 50 and obtain a percentile rank of 80. Now we know that this student has performed as well as or better than 80% of his/her peers. We would also say that this student is average. |
Readings
Chapter 17 Interpreting Test scores and Norms from Linn R.L. & Gronlund, N.E. (1995). Measurement and assessment in teaching. Englewood Cliffs, NJ: Merrill. |
|
|
|
|
|
|
|