Report
Detecting Rater Effects with a Multi-Faceted Rating Scale Model
Distributed by ERIC Clearinghouse
1997
Abstract
How common patterns of rater errors may be detected in a large-scale performance assessment setting is discussed. Common rater effects are identified, and a scaling method that can be used to detect them in operational data sets is presented. Simulated data sets are generated to exhibit each of these rater effects. The three continua that depict the most commonly cited rater effects are: (1) accuracy/randomness; (2) harshness/leniency; and (3) centrality/extremism. Rasch measurement theory provides one way of examining these rater effects within a normative framework. Rasch measurement places each facet of the measurement context on a common underlying linear scale, resulting in measures that can be subjected to traditional statistical analyses while allowing for unambiguous substantive interpretations of the meaning of examinee performance as it relates to rater performance and task functioning. In addition, Rasch calibrations of examinees, tasks, and raters are sample free in that they remove the influence of sample variability. The Multi-Faceted Rating Scale Model (MFRSM) of J. M. Linacre (1989) was used with simulated datasets that illustrate rater effects. Rater effects could be detected in the normative framework through MFRSM, and these effects seemed to operate on several continua. Further research is needed to determine how large a departure from the pool of raters needs to be before it can be detected in a normative framework. (Contains 5 figures, 9 tables, and 13 references.) (SLD)
Details
- Title: Subtitle
- Detecting Rater Effects with a Multi-Faceted Rating Scale Model
- Creators
- Edward W WolfeChris W. T Chiu
- Resource Type
- Report
- Publisher
- Distributed by ERIC Clearinghouse
- Number of pages
- 36 pages
- Language
- English
- Date published
- 1997
- Academic Unit
- Psychological and Quantitative Foundations
- Record Identifier
- 9985134747202771
Metrics
1 Record Views