Divesb.txt: 4 facet Interactions and Dummy Facets: Diving + Creativity

This is for 32-bit Facets 3.87. Here is Help for 64-bit Facets 4

Dummy facets are facets intended only for investigating interactions, not for measuring main effects. All the elements of a dummy facet are anchored at 0.

 

Example 1: At the 1988 Illinois Boys Diving Competition, 12 divers performed 3 dives which were rated by 7 judges. The divers chose which types of dive to perform, so each diver did not perform most types of dive. Not all ratings were recorded. Thus, most of the data is "missing"! The original ratings were in the range 0 to 10 with half point increments, so all ratings have been doubled to make them integers. Previous analyses have been performed so that good starting logit values are known. One diver did the wrong dive by mistake - this is deliberately treated as missing data in this analysis.

Include clearly aberrant data, e.g., lucky guesses, if your purpose is to describe the past, e.g., for awarding a prize. Omit such data if your purpose is to predict the future, e.g., to discover examinee's ability to perform surgery or to obtain item difficulty calibrations for an item bank.

 

We are investigating if there is any judge-by-round (= rater by time-point) interaction.

 

Facets specifications and data (in file Divesb.txt)

 

title = 1988 Illinois Boys Diving Competition (Anne Wendt)

facets = 4 ; four facets are diver, dives, round, judges

Inter-rater = 4 ; facet 4 is the rater facet

noncenter = 1 ; diver measures float

positive = 1 ; only for divers does greater score mean greater measure

models = ; these models will be scanned in order until a match is found for each datum:

 7,8,?,?,M ; make diver 7 on dive 8 missing - He did the wrong dive!

 ?,?,?B,?B,R20 ; "round"is a dummy facet for interaction with judges 

  ; but round is checked for validity

*

labels=

1,Diver  

;  diver          previous scores included in name for reference

 1=Marty Turek    292.85 425.65        , 2.08 ; logit starting value of 2.08 after second ","

|  ; 10 other divers

12=Bryan Hanania   251.15              ,-1.19

*

2,Dives  

 ; Each dives official weighted difficulty, e.g., 1.4, used as dive "name" after "="

1=1.4

| ; 6 other dives

8=2.6

9=2.4? ; unclear what dive this was

*

3,Round,A ; anchored at zero means dummy facet, does not affect measurement

1-3=,0 ; there were 3 rounds

*

4,Judges

1-7 ; there were 7 judges

*

data =

1,8,1,1,14 ; Diver 1, Marty, made a dive 8, a "2.6" dive, in round 1 rated by judge 1 with 7*2=14

| ; 136 other ratings

12,2,3,7,11 ; Bryan made a "1.7" dive in round 3 rated by judge 7 with 5.5*2=11

 

Largest judge by round interaction is:

 

Judge 1 was relatively lenient in Round 2 (+0.50 score points), but severe in Round 3 (-1.46 score points).

 

Example 2: In the Dives data, we suspect that there is an interaction between perceived diver ability (High and Low) and dive round. This is example dataset: Divesint.txt

 

Title = 1988 Illinois Boys Diving Competition (Anne Wendt)

Facets = 4 ; four facets are diver, dives, round, judges

Noncenter = 1 ; diver measures float

Positive = 1 ; only for divers does greater score mean greater measure

Models = ; these models will be scanned in order until a match is found for each datum:

 7,8,?,?,?,M ; make diver 7 on dive 8 missing - He did the wrong dive!

; interaction between facet 3 (round) and facet 5 (perceived ability)

 ?,?,?B,?,?B,DoublePoints ; Round excluded for this analysis by X 

*

Rating scale = DoublePoints,R20,Keep ; Keep unobserved intermediate categories

 0  = 0.0

 10 = 5.0 ; 5 on the original scoring

 20 = 10.0

*

labels=

1,Diver

;  diver          previous scores

1, H: Marty Turek     292.85  ; previous scores

2, H: Tom Wright      279.95  

3, L: Mike Gotkowski  249.9

4, L: Matt Paulson    244.55

5, L: Scott Ternovits 252.8

6, L: Ross Moyer      243.4 

7, H: Curt Billings   266.25 

8, H: Steve Hutchings 267.15

9, L: Larry Kirk      258.35

10, H: Kurt Becker     284.4

11, L: Lance Kleffman  259.6

12, L: Bryan Hanania   251.15

*

2,Dives

1-9

*

3,Round, D ; a dummy facet

1-3

*

4,Judges

1-7

*

5, Perceived ability, D ; a dummy facet

1, High previous score 260 up

2, Low previous score 259 down

*

Dvalues =

4, 1-7 ; the observations for the 7 judges are in Excel

5, 1, 1, 1 ; element label for facet 5 is first character of label in facet 1

*

Data = Dives.xls 

; has this format:

; diver

dive

round

judge 1

judge 2

judge 3

judge 4

judge 5

judge 6

judge 7

1

8

1

14

11

12

12

.

.

.

1

1

2

16

13

13

16

14

13

14

2

8

2

12

12

10

11

10

11

13

 

The interactions indicate a jump in ability in round 3 for those with perceived high ability. Do they handle the pressure of the final round better? Are the judges more lenient with them in the third round? Or ... This could be the start of an exciting investigation into competitive behavior of great interest to diving coaches.

 

 

Example 3: We want to investigate whether there is an interaction between judges and score-bands for the Creativity (Guilford) data. A dummy facet "score-band" is introduced into the data. It has 3 elements: 1, 4, 7, corresponding to the observed rating ranges 1-3, 4-6, 7-9.

 

Facets specifications and data (in file Guilfordsb.txt):

 

; Guilfordsb.txt

Facets = 4 ; four facets judges, examinees, items, score-band

Inter-rater = 1 ; facet 1 is the rater facet

Model = ?B,?,?,?B, R9 ; interaction between judges and score-band

 

Labels=               ;to name the components

1,Senior scientists   ;name of first facet raters

1=Avogadro            ;names of elements within facet

2=Brahe               ;these must be named, or they are treated as missing

3=Cavendish

*

2,Junior Scientists

2=Betty

5=Edward

7=George

1=Anne  

3=Chris

4=David

6=Fred

*

3,Traits

1=Attack

2=Basis

3=Clarity

4=Daring

5=Enthusiasm

*

4, Score-band, A  ; anchored

1, Low 1-3, 0     ; 1+2+3 ; anchored at 0 to avoid distorting the measures

4, Medium 4-6, 0  ; 4+5+6  

7, High 7-9, 0    ; 7+8+9

*

 

; rater, examinee, item, score band, response

Data=  

1,1,1,4,5

1,1,2,4,5

....


Help for Facets Rasch Measurement and Rasch Analysis Software: www.winsteps.com Author: John Michael Linacre.
 

Facets Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Minifac download
Winsteps Rasch measurement software. Buy for $149. & site licenses. Freeware student/evaluation Ministep download

Rasch Books and Publications: Winsteps and Facets
Applying the Rasch Model (Winsteps, Facets) 4th Ed., Bond, Yan, Heene Advances in Rasch Analyses in the Human Sciences (Winsteps, Facets) 1st Ed., Boone, Staver Advances in Applications of Rasch Measurement in Science Education, X. Liu & W. J. Boone Rasch Analysis in the Human Sciences (Winsteps) Boone, Staver, Yale Appliquer le modèle de Rasch: Défis et pistes de solution (Winsteps) E. Dionne, S. Béland
Introduction to Many-Facet Rasch Measurement (Facets), Thomas Eckes Rasch Models for Solving Measurement Problems (Facets), George Engelhard, Jr. & Jue Wang Statistical Analyses for Language Testers (Facets), Rita Green Invariant Measurement with Raters and Rating Scales: Rasch Models for Rater-Mediated Assessments (Facets), George Engelhard, Jr. & Stefanie Wind Aplicação do Modelo de Rasch (Português), de Bond, Trevor G., Fox, Christine M
Exploring Rating Scale Functioning for Survey Research (R, Facets), Stefanie Wind Rasch Measurement: Applications, Khine Winsteps Tutorials - free
Facets Tutorials - free
Many-Facet Rasch Measurement (Facets) - free, J.M. Linacre Fairness, Justice and Language Assessment (Winsteps, Facets), McNamara, Knoch, Fan
Other Rasch-Related Resources: Rasch Measurement YouTube Channel
Rasch Measurement Transactions & Rasch Measurement research papers - free An Introduction to the Rasch Model with Examples in R (eRm, etc.), Debelak, Strobl, Zeigenfuse Rasch Measurement Theory Analysis in R, Wind, Hua Applying the Rasch Model in Social Sciences Using R, Lamprianou Journal of Applied Measurement
Rasch Models: Foundations, Recent Developments, and Applications, Fischer & Molenaar Probabilistic Models for Some Intelligence and Attainment Tests, Georg Rasch Rasch Models for Measurement, David Andrich Constructing Measures, Mark Wilson Best Test Design - free, Wright & Stone
Rating Scale Analysis - free, Wright & Masters
Virtual Standard Setting: Setting Cut Scores, Charalambos Kollias Diseño de Mejores Pruebas - free, Spanish Best Test Design A Course in Rasch Measurement Theory, Andrich, Marais Rasch Models in Health, Christensen, Kreiner, Mesba Multivariate and Mixture Distribution Rasch Models, von Davier, Carstensen
As an Amazon Associate I earn from qualifying purchases. This does not change what you pay.

facebook Forum: Rasch Measurement Forum to discuss any Rasch-related topic

To receive News Emails about Winsteps and Facets by subscribing to the Winsteps.com email list,
enter your email address here:

I want to Subscribe: & click below
I want to Unsubscribe: & click below

Please set your SPAM filter to accept emails from Winsteps.com
The Winsteps.com email list is only used to email information about Winsteps, Facets and associated Rasch Measurement activities. Your email address is not shared with third-parties. Every email sent from the list includes the option to unsubscribe.

Questions, Suggestions? Want to update Winsteps or Facets? Please email Mike Linacre, author of Winsteps mike@winsteps.com


State-of-the-art : single-user and site licenses : free student/evaluation versions : download immediately : instructional PDFs : user forum : assistance by email : bugs fixed fast : free update eligibility : backwards compatible : money back if not satisfied
 
Rasch, Winsteps, Facets online Tutorials

Coming Rasch-related Events
May 17 - June 21, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 12 - 14, 2024, Wed.-Fri. 1st Scandinavian Applied Measurement Conference, Kristianstad University, Kristianstad, Sweden http://www.hkr.se/samc2024
June 21 - July 19, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Winsteps), www.statistics.com
Aug. 5 - Aug. 6, 2024, Fri.-Fri. 2024 Inaugural Conference of the Society for the Study of Measurement (Berkeley, CA), Call for Proposals
Aug. 9 - Sept. 6, 2024, Fri.-Fri. On-line workshop: Many-Facet Rasch Measurement (E. Smith, Facets), www.statistics.com
Oct. 4 - Nov. 8, 2024, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
Jan. 17 - Feb. 21, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
May 16 - June 20, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com
June 20 - July 18, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Further Topics (E. Smith, Facets), www.statistics.com
Oct. 3 - Nov. 7, 2025, Fri.-Fri. On-line workshop: Rasch Measurement - Core Topics (E. Smith, Winsteps), www.statistics.com

 

Our current URL is www.winsteps.com

Winsteps® is a registered trademark