Displacement measures

DISPLACE column should only appear with anchored or TARGET= runs. Otherwise its appearance indicates lack of convergence.  If small displacements are being shown, try tightening the convergence criteria, LCONV=.

 

Anchored analyses, IAFILE=, PAFILE=: if large displacements are shown for the anchored items or persons, try changing the setting of ANCESTIM=.

 

The displacement is an estimate of the amount to add to the MEASURE to make it conform with the data.

 

Positive displacement for a person ability indicates that the observed person score is higher than the expected person score based on the reported measure (usually an anchor value).

 

Positive displacement for an item difficulty indicates that the observed item score is lower than the expected item score based on the reported measure (usually an anchor value).

 


 

Anchored item difficulties with Displacements: what Item Difficulty for next time?

 

1.Do the analysis that produces the displacements with IAFILE= (and SAFILE=). Output PFILE=pf.txt.

2.Treat these person abilities as "true", especially if you are reporting them.

3.Reanalyze the data with PAFILE=pf.txt and IAFILE=nothing (and SAFILE=nothing). This will produce an IFILE= (and SFILE=) with no displacements for which the item difficulties are the best match to the anchored person abilities.

4.You can compare the IFILE= from 3 with the (difficulty and displacement)  from 1. The values should be close, but the IFILE= values from 3. will produce the same person abilities from the same data next time as they do this time. (difficulty and displacement) will produce different sets of person abilities in this situation.

 


Displacement Computation

 

The displacement for each item or person, with all the items, persons, thresholds anchored at their final values.

 

The computation uses iterative curve-fitting. For an item (or similarly a person)

1.Compute the expected score for (final item difficulty= D0) = E0

2.Compute the expected score for (final item difficulty+1 logit = D1) = E1

3.Points (D0,E0) and (D1,E1) are on a logistic score-to-measure ogive:

4.the observed score for this item is S. The matching point on the logistic ogive is DS.

5.the displacement is (DS-D0).

 

Since we do not want to recalibrate all the items, we are only worried about items (persons) that have drifted too far from their difficulties in the anchor set. These are items with big displacements.

 

If there is a high proportion of drifted items (or even 1 drifted item in high stakes situations), then apply this method. The change using this method will be bigger than the displacement because the drifted item has influenced all the other measures.

 

1)Reanalyze the data with all other items anchored, unanchoring the suspect item(s) and giving them a weight of zero (IWEIGHT=), and unanchoring the persons.

2)The persons get new measures based on the good anchored items

3)The suspect items get difficulties based on the new person measures.

4)The zero weights of the items allows them to be estimated from the new person measures without influencing the item or person measures.

5)Drift = estimated value - anchored value

 

Faster computation with complete data:

Compare the anchor value with the measure for the observed score in Table 20.3 (items) or Table 20.1 (persons)

 


 

The DISPLACE value is the size of the change in the parameter estimate that would be observed in the next estimation iteration if this parameter was free (unanchored) and all other parameter estimates were anchored at their current values.

 

For a parameter (item or person) that is anchored in the main estimation, DISPLACE indicates the size of disagreement between an estimate based on the current data and the anchor value.

 

For an unanchored item, if the DISPLACE value is large enough to be of concern, then the convergence criteria are not tight enough LCONV=, RCONV=, CONVERGE=, MJMLE=

 

DISPLACE approximates the displacement of the estimate away from the statistically better value which would result from the best fit of your data to the model. Each DISPLACE value is computed as though all other parameter estimates are exact. Only meaningfully large values are displayed. They indicate lack of convergence, or the presence of anchored or targeted values. The best fit value can be approximated by adding the displacement to the reported measure or calibration.

 

Note: Winsteps uses a logistic-curve-fitting approach to estimation. This has proved more robust than Newton-Raphson for badly-behaved data (long rating scales with missing categories, sparse datasets, etc.). Essentially, we know that all the underlying functions in Rasch estimation are logistic ogives, so we can take advantage of this to predict the next (improved) estimate based on the current estimate, its expected score, and the observed score. For computing displacements, logistic-curve-fitting can be replicated in Excel. You have the expected scores for the persons on the anchored item from the XFILE=. Then compute the expected scores for the person on the anchored item difficulty + 1 logit using the person and item measures in the XFILE=. This gives two total expected item scores. For each total expected item score and the total  observed item score, compute the log odds: ln (value/(max possible item score - value)). This gives 3 log odds: E0, E1 and O. Then displacement is (O - E0)/(E1 - E0) logits. This value of the displacement is usually very close to the Newton-Raphson value.

 

Standard Error of the Displacement Measure

+----------------------------------------------------------------------------------------+

|ENTRY    RAW                   MODEL|   INFIT  |  OUTFIT  |PTMEA|        |              |

|NUMBER  SCORE  COUNT  MEASURE  S.E. |MNSQ  ZSTD|MNSQ  ZSTD|CORR.|DISPLACE| TAP          |

|------------------------------------+----------+----------+-----+--------+--------------|

|     3     35     35    2.00A    .74| .69   -.6| .22    .5|  .00|   -3.90| 1-2-4        |

 

Since the reported "measure" is treated as a constant when "displacement" is computed, the S.E. of the reported "measure" actually is the same as the S.E. of the displacement. The DISPLACE column shows the displacement in the same units as the MEASURE. This is logits when USCALE=1, the default. If the anchored measure value is considered to be exact, i.e., a point-estimate, then the S.E. standard error column indicates the standard error of the displacement. The statistical significance of the Displacement is given by

t = DISPLACE / S.E. with approximately COUNT degrees of freedom.

 

This evaluates how likely the reported size of the displacement is, if its "true" size is zero. But both the displacements and their standard errors are estimates, so the t-value may be slightly mis-estimated. Consequently allow for a margin of error when interpreting the t-values.

 

If the anchored measure value has a standard error obtained from a different data set, then the standard error of the displacement is:

S.E. (Displacement) = Sqrt(S.E.² + S.E.²(anchor value from original data) )

 

When does large displacement indicate that an item or person should be unanchored or omitted?

This depends on your purpose. If you are anchoring items in order to measure three additional people to add to your measured database of thousands, then item displacement doesn't matter.

 

Anchor values should be validated before they are used. Do two analyses:

(a) with no items anchored (i.e., all items floating), produce person and item measures.

(b) with anchored items anchored, produce person and item measures.

 

Then cross-plot the item difficulties for the two runs, and also the person measures. The person measures will usually form an almost straight line.

 

For the item difficulties, unanchored items will form a straight-line. Some anchored items may be noticeably off the line. These are candidates for dropping as anchors. The effect of dropping or un-anchoring a "displaced" anchor item is to realign the person measures by roughly (displacement / (number of remaining anchored items)).

 

Random displacements of less than 0.5 logits are unlikely to have much impact in a test instrument.

"In other work we have found that when [test length] is greater than 20, random values of [discrepancies in item calibration] as high as 0.50 [logits] have negligible effects on measurement." ( Wright & Douglas, 1976, "Rasch Item Analysis by Hand")

 

"They allow the test designer to incur item discrepancies, that is item calibration errors, as large as 1.0 [logit]. This may appear unnecessarily generous, since it permits use of an item of difficulty 2.0, say, when the design calls for 1.0, but it is offered as an upper limit because we found a large area of the test design domain to be exceptionally robust with respect to independent item discrepancies." (Wright & Douglas, 1975, "Best Test Design and Self-Tailored Testing.")

 

Most DIF work seems to be done by statisticians with little interest in, and often no access to, the substantive material. So they have no qualitative criteria on which to base their DIF acceptance/rejection decisions. The result is that the number of items with DIF is grossly over-reported (Hills J.R. (1989) Screening for potentially biased items in testing programs. Educational Measurement: Issues and practice. 8(4) pp. 5-11).


Help for Winsteps Rasch Measurement and Rasch Analysis Software: www.winsteps.com. Author: John Michael Linacre

Facets Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Minifac download
Winsteps Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Ministep download

Rasch Books and Publications
Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences, 2nd Edn, 2024 George Engelhard, Jr. & Jue Wang Applying the Rasch Model (Winsteps, Facets) 4th Ed., Bond, Yan, Heene Advances in Rasch Analyses in the Human Sciences (Winsteps, Facets) 1st Ed., Boone, Staver Advances in Applications of Rasch Measurement in Science Education, X. Liu & W. J. Boone Rasch Analysis in the Human Sciences (Winsteps) Boone, Staver, Yale
Introduction to Many-Facet Rasch Measurement (Facets), Thomas Eckes Statistical Analyses for Language Testers (Facets), Rita Green Invariant Measurement with Raters and Rating Scales: Rasch Models for Rater-Mediated Assessments (Facets), George Engelhard, Jr. & Stefanie Wind Aplicação do Modelo de Rasch (Português), de Bond, Trevor G., Fox, Christine M Appliquer le modèle de Rasch: Défis et pistes de solution (Winsteps) E. Dionne, S. Béland
Exploring Rating Scale Functioning for Survey Research (R, Facets), Stefanie Wind Rasch Measurement: Applications, Khine Winsteps Tutorials - free
Facets Tutorials - free
Many-Facet Rasch Measurement (Facets) - free, J.M. Linacre Fairness, Justice and Language Assessment (Winsteps, Facets), McNamara, Knoch, Fan
Other Rasch-Related Resources: Rasch Measurement YouTube Channel
Rasch Measurement Transactions & Rasch Measurement research papers - free An Introduction to the Rasch Model with Examples in R (eRm, etc.), Debelak, Strobl, Zeigenfuse Rasch Measurement Theory Analysis in R, Wind, Hua Applying the Rasch Model in Social Sciences Using R, Lamprianou El modelo métrico de Rasch: Fundamentación, implementación e interpretación de la medida en ciencias sociales (Spanish Edition), Manuel González-Montesinos M.
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Rasch Models for Measurement, David Andrich Constructing Measures, Mark Wilson Best Test Design - free, Wright & Stone
Rating Scale Analysis - free, Wright & Masters
Virtual Standard Setting: Setting Cut Scores, Charalambos Kollias Diseño de Mejores Pruebas - free, Spanish Best Test Design A Course in Rasch Measurement Theory, Andrich, Marais Rasch Models in Health, Christensen, Kreiner, Mesba Multivariate and Mixture Distribution Rasch Models, von Davier, Carstensen
As an Amazon Associate I earn from qualifying purchases. This does not change what you pay.

facebook Forum: Rasch Measurement Forum to discuss any Rasch-related topic

To receive News Emails about Winsteps and Facets by subscribing to the Winsteps.com email list,
enter your email address here:

I want to Subscribe: & click below
I want to Unsubscribe: & click below

Please set your SPAM filter to accept emails from Winsteps.com
The Winsteps.com email list is only used to email information about Winsteps, Facets and associated Rasch Measurement activities. Your email address is not shared with third-parties. Every email sent from the list includes the option to unsubscribe.

Questions, Suggestions? Want to update Winsteps or Facets? Please email Mike Linacre, author of Winsteps mike@winsteps.com


State-of-the-art : single-user and site licenses : free student/evaluation versions : download immediately : instructional PDFs : user forum : assistance by email : bugs fixed fast : free update eligibility : backwards compatible : money back if not satisfied
 
Rasch, Winsteps, Facets online Tutorials


 

 
Coming Rasch-related Events: Winsteps and Facets
Oct 21 - 22 2024, Mon.-Tues. In person workshop: Facets and Winsteps in expert judgement test validity - UNAM (México) y Universidad Católica de Colombia. capardo@ucatolica.edu.co, benildegar@gmail.com
Oct. 4 - Nov. 8, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 17 - Feb. 21, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
May 16 - June 20, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 20 - July 18, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Facets), www.statistics.com
Oct. 3 - Nov. 7, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com

 

 

Our current URL is www.winsteps.com

Winsteps® is a registered trademark