Bias interaction DIF DPF DRF estimation

This is for 32-bit Facets 3.87. Here is Help for 64-bit Facets 4

1. If all facets are positive, higher raw score = more person ability. Higher raw score = more rater leniency. Higher raw score = more task easiness.

Positive bias/interaction means that the observed score is higher than expected.

Suppose we are modeling task-by-rater interaction, and the interaction is positive (score higher than expected)

Then if we think that the interaction is caused by the task, then the "absolute measure" of the task = task easiness + interaction.  "Absolute measure" of the rater = rater leniency.

Then if we think that the interaction is caused by the rater, then the "absolute measure" of the task = task easiness.  "Absolute measure" of the rater = rater leniency + interaction.

 

2. If some facets are positive and some facets are negative, then the relationship between raw scores, measures and interactions becomes complicated. Plots and reports differ depending on whether we model facets to be positive or negative, interactions to be positive or negative and to which facet we want to add the interaction. Example:

Positive: higher raw score = more person ability.

Positive; Higher raw score = more rater leniency.

Negative: Higher raw score = less task difficulty.

We must choose: Positive bias/interaction means that the score is higher than expected or lower than expected.

If we think of interaction for the rater, then we want more score = positive (more lenient) which add to rater leniency. If we think of interaction for the task, then we want more score = negative (more difficult) which adds to task difficulty making the task less difficult.

This becomes confusing. So I strongly recommend that all facets are positive if interactions can apply to both positive and negative facets.

 


 

After estimating the measures, Facets checks to see if any Model= specifications include Bias specifiers, "B". If so, for each such model, the specified Bias interaction is estimated for all the data (not just the data matching that particular model). Bias can be due to any type of interaction including Differential Item Functioning DIF, Differential Person Functioning DPF, Differential Rater Functioning DRF.

This is done by iterating through the data again and, after convergence, doing one further iteration to calculate statistics.

 

Computation of interactions is a two-stage process.

1. The measures for the elements, and the structure of the rating scale, are estimated. Then those values are anchored (fixed, held constant).

2. The expected values of the observations are subtracted from the observed values of the observations, producing the residuals.

3. The residuals corresponding to each interaction term (e.g., examinees rated by judge 4) are summed. If this sum is not zero, then there is an interaction.

4. The size of the interaction is estimated. A first approximation is:

Interaction (logits) = (sum of residuals) / (sum of the statistical-information in the observations).

 

Algebraically, first the Bn, Di, Cj, Fk are estimated using a Rasch model such as:

 

log ( Pnijk / Pnij(k-1)) = Bn - Di - Cj - Fk

 

Then the Bn, Di, Cj, Fk are anchored, and the bias/interaction terms, e.g., Cij, are estimated:

 

log ( Pnijk / Pnij(k-1)) = ( Bn - Di - Cj - Fk ) - Cij

 

Thus the Cij are estimated from the residuals left over from the main analysis. The conversion from residual score to bias interaction size is non-linear. Bias sizes may not sum to zero.

 

Bias, (also called interaction, differential item function, differential person function, etc.,) estimation serves several purposes:

 

1) in diagnosing misfit:

The response residuals are partitioned by element, e.g., by judge-item pairs, and converted into a logit measure. Estimates of unexpected size and statistical significance flag systematic misfit, focusing the misfit investigation.

 

2) in investigating validity:

A systematic, but small, bias in an item or a judge, for or against any group of persons, may be overwhelmed by the general stochastic component in the responses. Consequently it may not be detected by the usual summary fit statistics. Specifying a bias analysis between elements of facets of particular importance provides a powerful means of investigating and verifying the fairness and functioning of a test.

 

3) in assessing the effect of bias:

Since bias terms have a measure and a standard error (precision), their size and significance (t-statistic) are reported. This permits the effect of bias to be expressed in the same frame of reference as the element measures. Thus each element measure can be adjusted for any bias which has affected its estimation, e.g., by adding the estimate of bias, which has adversely affected an element, to that element's logit measure. Then the practical implications of removing bias can be determined. Does adjustment for bias alter the pass-fail decision? Does adjustment for bias affect the relative performance of two groups in a meaningful way?

 

4) in partitioning unexplained "error" variance:

The bias logit sample standard deviation corrected for its measurement error, can be an estimate of the amount of systematic error in the error variance (RMSE).

e.g., for a bias analysis of judges,

Bias logit S.D. = 0.47, mean bias S.E. = 0.32 (Table 13),

so "true" bias S.D. = (0.47² - 0.32²) = 0.35 logits,

but, this exceeds the RMSE for judges = 0.12 (Table 7).

Here, locally extreme judge-person scores cause an overestimation of systematic bias.

 


 

Adjusting for bias:

 

A straight-forward approach is to define the biased element as two elements: one element for one subset of judges (examinees, etc.) and a different items for the other subset of judges (examinees, etc.). This can done by defining an extra item element, and then adjusting item references in the data file accordingly.

 

Example:

 

Facets = 4 ; Items, candidates, examiners, bias adjustment

Non-center = 2 ; candidates float

 

Models =

?, 28, 17, 1, myscale ; allow for bias adjustment between candidate 28 and examiner 17

?,  ?, ?, 2, myscale

*

 

Rating scale = myscale, R9

 

Labels=

1, Items

...

2, Candidates

...

3, Examiners

....

4, Bias adjustment, A

1, 28-17 adjustment ; the bias will be absorbed by this element, relative to element 2.

2, Everyone else, 0

*

 

Data=

1-5, 28, 17, 1, 1,2,3,4,5

1-5, 29, 23, 2, 5,4,3,2,1

.....

 


 

Non-Uniform DIF

 

Create a dummy facet for ability levels: then a three-way interaction between

person group dummy facet, ability level dummy facet, and item (or whatever).

Models=?,?B,?B,?,?B,R3   (or whatever)

 


 

Paired-Comparison Bias/Interaction Analysis:

 

Facets produces meaningful numbers in the Bias/interaction analysis when:

1) Use mirrored data, but set the weight = 1.0, instead of 0.5. For the main analysis, use weight 0.5.

2) Arrange the data so that the Models= is  ..., -?,?,... instead of ...,?,-?,...


Help for Facets Rasch Measurement and Rasch Analysis Software: www.winsteps.com Author: John Michael Linacre.
 

Facets Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Minifac download
Winsteps Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Ministep download

Rasch Books and Publications: Winsteps and Facets
Applying the Rasch Model (Winsteps, Facets) 4th Ed., Bond, Yan, Heene Advances in Rasch Analyses in the Human Sciences (Winsteps, Facets) 1st Ed., Boone, Staver Advances in Applications of Rasch Measurement in Science Education, X. Liu & W. J. Boone Rasch Analysis in the Human Sciences (Winsteps) Boone, Staver, Yale Appliquer le modèle de Rasch: Défis et pistes de solution (Winsteps) E. Dionne, S. Béland
Introduction to Many-Facet Rasch Measurement (Facets), Thomas Eckes Rasch Models for Solving Measurement Problems (Facets), George Engelhard, Jr. & Jue Wang Statistical Analyses for Language Testers (Facets), Rita Green Invariant Measurement with Raters and Rating Scales: Rasch Models for Rater-Mediated Assessments (Facets), George Engelhard, Jr. & Stefanie Wind Aplicação do Modelo de Rasch (Português), de Bond, Trevor G., Fox, Christine M
Exploring Rating Scale Functioning for Survey Research (R, Facets), Stefanie Wind Rasch Measurement: Applications, Khine Winsteps Tutorials - free
Facets Tutorials - free
Many-Facet Rasch Measurement (Facets) - free, J.M. Linacre Fairness, Justice and Language Assessment (Winsteps, Facets), McNamara, Knoch, Fan
Other Rasch-Related Resources: Rasch Measurement YouTube Channel
Rasch Measurement Transactions & Rasch Measurement research papers - free An Introduction to the Rasch Model with Examples in R (eRm, etc.), Debelak, Strobl, Zeigenfuse Rasch Measurement Theory Analysis in R, Wind, Hua Applying the Rasch Model in Social Sciences Using R, Lamprianou Journal of Applied Measurement
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Rasch Models for Measurement, David Andrich Constructing Measures, Mark Wilson Best Test Design - free, Wright & Stone
Rating Scale Analysis - free, Wright & Masters
Virtual Standard Setting: Setting Cut Scores, Charalambos Kollias Diseño de Mejores Pruebas - free, Spanish Best Test Design A Course in Rasch Measurement Theory, Andrich, Marais Rasch Models in Health, Christensen, Kreiner, Mesba Multivariate and Mixture Distribution Rasch Models, von Davier, Carstensen
As an Amazon Associate I earn from qualifying purchases. This does not change what you pay.

facebook Forum: Rasch Measurement Forum to discuss any Rasch-related topic

To receive News Emails about Winsteps and Facets by subscribing to the Winsteps.com email list,
enter your email address here:

I want to Subscribe: & click below
I want to Unsubscribe: & click below

Please set your SPAM filter to accept emails from Winsteps.com
The Winsteps.com email list is only used to email information about Winsteps, Facets and associated Rasch Measurement activities. Your email address is not shared with third-parties. Every email sent from the list includes the option to unsubscribe.

Questions, Suggestions? Want to update Winsteps or Facets? Please email Mike Linacre, author of Winsteps mike@winsteps.com


State-of-the-art : single-user and site licenses : free student/evaluation versions : download immediately : instructional PDFs : user forum : assistance by email : bugs fixed fast : free update eligibility : backwards compatible : money back if not satisfied
 
Rasch, Winsteps, Facets online Tutorials

Coming Rasch-related Events
May 17 - June 21, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 12 - 14, 2024, Wed.-Fri. 1st Scandinavian Applied Measurement Conference, Kristianstad University, Kristianstad, Sweden http://www.hkr.se/samc2024
June 21 - July 19, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com
Aug. 5 - Aug. 6, 2024, Fri.-Fri. 2024 Inaugural Conference of the Society for the Study of Measurement (Berkeley, CA), Call for Proposals
Aug. 9 - Sept. 6, 2024, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com
Oct. 4 - Nov. 8, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 17 - Feb. 21, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
May 16 - June 20, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 20 - July 18, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Facets), www.statistics.com
Oct. 3 - Nov. 7, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com

 

Our current URL is www.winsteps.com

Winsteps® is a registered trademark