FAQ/icc - CBU statistics Wiki

Revision 32 as of 2010-11-17 12:30:11

Clear message
location: FAQ / icc

Intraclass correlations

An alternative, to the kappa statistic, which uses an analysis of variance output to estimate rater reliability is the intraclass correlation coefficient (ICC).

For a repeated measures anova involving k raters it follows assuming both subjects and raters are fixed effects that

$$\mbox{MS(total w)} = \frac{\mbox{SS(raters)–MS(subjects x raters)}}{\mbox{df(raters) + df(subjects x raters)}}$$

$$\mbox{ICC1} = \frac{\mbox{MS(subjects)–MS(total w)}}{\mbox{MS(subjects) + (k-1)MS(total w)}}$$

where MS is the Mean square from the repeated measures analysis of variance.

This corresponds to the one-way random approach in SPSS (see [http://www.uvm.edu/~dhowell/StatPages/More_Stuff/icc/icc.html here).]

It follows that the intra-class correlation (ICC), unlike the Pearson correlation, is useful for pooling paired data each having three or more observations. Einfield and Tonge (1992, p 12) prefer using the ICC to the Pearson as it is more conservative owing to that fact it "takes account of the absolute as well as the relative difference between the scores of two raters".

Howell (1997) also recommends an alternative, most widely used ICC which assumes that the raters are a random sample from a larger population which has an extra term in the denominator and is of form

$$\mbox{ICC2} = \frac{\mbox{MS(subjects)–MS(subjects x raters)}}{\mbox{MS(subjects) + (k-1)MS(subjects x raters) + k[MS(raters) - MS(subjects x raters)]/n}}$$

where n is the number of subjects being rated.

e.g. if two raters rate three subjects giving ratings 1,2; 2,4; 3,6 then ICC1 = 0.80 and ICC2 = 0.46.

ICC may be computed in SPSS using analyze>scale>reliability analysis>statistics and choosing one of the two ICCs which allow a type of absolute agreement.

Examples of ICC computation in SPSS [http://www.nyu.edu/its/statistics/Docs/intracls.html are available here] and [attachment:ICC.doc here.] The fixed ICC correlations called sfsingle, sf random and sffixed in the above article are of form

$$\frac{\mbox{true inter-rater variance}}{\mbox{true inter-rater variance + common error in rating variance}}$$

as mentioned as a reliability correlation in the two rater case, for example, in [http://www-users.york.ac.uk/%7Emb55/talks/oxtalk.htm a paper by Martin Bland and Doug Altman.]

An overview of approaches to inter rater reliability including the ICC is given by Darroch and McCloud (1986).

  • [:FAQ/iccpr: Inferiority of using a Pearson correlation compared to an ICC]

References

Darroch JN, McCloud PI (1986) Category distinguishability and observer agreement Australian Journal of Statistics 28 371-88.

Howell DC (1997) Statistical methods for psychologists. Fourth edition. Wadsworth:Belmont,CA. (pages 490-493).

Einfield, SL and Tonge, BJ (1992) Manual for the developmental hebaviour checklist (DBC)(Primary Carer version). Melbourne:School of Psychiatry, Unievrsity of new South Wales, and Centre for Developmental Psychiatry, Monash University, Clayton, Victoria.

Shrout, PE and Fleiss, JL (1979). Intraclass Correlations: Uses in Assessing Rater Reliability, Psychological Bulletin, 86 (2) 420-428. (A good primer showing how anova output can be used to compute ICCs).