box scores and kappa…

Posted on Tuesday 8 May 2012

Lou Gehrigs Box ScoresMy Dad had been a ball player, but by the time I came along, he was a Coach – my uncle too. People on the street called them both ‘Coach’ and it sounded like ‘Doctor’ or ‘Mr. President’ and I felt proud. They had grown up poor. Sports were a way to escape working in the mines, so it had to be a consuming interest for it to succeed – and it was. I grew up with sports and the Stats as the most important things going. The first part of the newspaper my Dad read were the Box-Scores from last night’s games in whatever sport was in season, in preparation for the evening game on the radio. I remember when television came and you could actually see the game, it was different. I was used to seeing the game in fantasy with pictures from the baseball cards, sports page, and stats from the box scores all dancing in my head. Being able to actually see the game played took some getting used to.

Babe Ruth, Shoeless Joe JacksonThere were live games too at the local park [Coach’s kids got in free]. In the stands, there were lots of people who scored the game on an elaborate form in the program. They kept up with every pitch, every hit, and drew pictures with special symbols of the play. They were the ones who knew their Stats and would predict things [like in the recent movie Moneyball]. It was like the players weren’t real people. They were Stats too – ERAs, RBIs, Steals. "He leads the league in whatever"

When I got to Statistics in school, it was mostly about probability – how probable is it that something is significant, or is it just chance. I had to get used to that. In Baseball, everything had been significant. This was a different kind of Stats. Even in a Master’s program in Statistics back when, it was mostly about probability – a kind of yes-no science. With Baseball, there was so much more. That Babe Ruth or Shoeless Joe were probably going to get hits wasn’t enough. What kind of hit was the question. A double? a triple? a homer? an RBI? Could you count on them in a clinch? It was more than probability, it was effect size [how strong was Babe’s mojo?] and reliability [how often did Shoeless Joe get on base? steal?]. I’m just learning about those kinds of Stats in medicine – things like NNT [number needed to treat], OR [odds ratio], Effect Size, and Kappa. And Kappa is on the front burner right now since the DSM-5 Field Trials have been completed. Kappa is about reliability.

Robert Spitzer was not well suited for the psychoanalytic training he chose. Psychoanalysis and psychotherapy are about intuition, and by report, Spitzer was remarkably short on intuition, so he was a perfect person to focus on the concrete descriptive criteria for psychiatric diagnosis. Along the way, he teamed up with Statisticians Joseph Fleiss of Columbia and Jacob Cohen at NYU who had studied inter-rater reliability, and began to use their versions of Kappa, a statistical method for quantification:
Robert Spitzer [and the ghost of Emil Kraepelin] 
A Re-analysis of the Reliability of Psychiatric Diagnosis
By ROBERT L. SPITZER and JOSEPH L. FLEISS
British Journal of Psychiatry. 1974  125:341-347.

Classification systems such as diagnosis have two primary properties, reliability and validity. Reliability refers to the consistency with which subjects are classified; validity, to the utility of the system for its various purposes. In the case of psychiatric diagnosis, the purposes of the classification system are communication about clinical features, aetiology, course of illness, and treatment.  A necessary constraint on the validity of a system is its reliability. There is no guarantee that a reliable system is valid, but assuredly an unreliable system must be invalid. Studies of the reliability of psychiatric diagnosis provide information on the upper limits of its validity. This paper discusses some of the difficulties in appraising diagnostic reliability, offers a re-analysis of available data from the literature, and suggests a possible course of action to improve psychiatric diagnosis…
That "course of action" was to use Kappa to test agreement of raters using given same criteria, and thus the nidus of the DSM-III was in place. At its core, Kappa quantifies the interval between what pure chance would bring two raters and total agreement between them:
 
So if "chance" agreement was 50:50 and they actually agreed on 3 out of 4, kappa would equal (0.75 – 0.50)÷(1.00 – 0.50) = 0.50. Things get more complex when you throw in multiple raters and multiple categories with varying prevalence, but the concept remains the same – quantifying the interval between chance [kappa = 0] and perfect agreement [kappa = 1]. In his 1974 article, Spitzer said [self-evident…]:
There are no diagnostic categories for which reliability is uniformly high. Reliability appears to be only satisfactory for three categories: mental deficiency, organic brain syndrome [but not its subtypes], and alcoholism. The level of reliability is no better than fair for psychosis and schizophrenia and is poor for the remaining categories.
So here’s his original table color coded by yours truly based on what he said in the article:
[I assume that Chronic Brain Syndrome was a slip-up] Spitzer’s criteria were pretty clear. And this is Robert Spitzer’s show – he invented it in collaboration with the very statisticians that created kappa. In the article, he describes the conditions for each of the studies he reports above. All were rated by psychiatrists. Some were tightly controlled, but others were rated by admitting psychiatry  residents. Spitzer concludes:
The reliability of psychiatric diagnosis as it has been practised since at least the late 1950’s is not good. It is likely that the reasons for diagnostic unreliability are the same now as when Beck et al studied them. They found that a significant amount of the variability among diagnosticians was due to differences in how they elicited and evaluated the necessary information, and that an even larger amount was due to inherent weakness and ambiguities in the nomenclature. Since that time there have been two major innovations which may provide solutions to these problems.

Several investigators have developed structured interview schedules which an interviewer uses in his examination of the patient. These techniques provide for a standardized sequence of topics, and ensure that variability among clinicians in how they conduct their interviews and in what topics they cover is kept to a minimum. For rating the pathology observed, these schedules contain precoded items which explicitly define the behaviours to be rated rather than relying on technical terms which have different meanings to different clinicians.

With respect to improving the nomenclature, the St.Louis group (Feighner et al) has offered a system limited to 16 diagnoses for which they believe strong validity evidence exists, and for which specified requirements are provided. Whereas in the standard system the clinician determines to which of the various diagnostic stereotypes his patient is closest, in the St. Louis system the clinician determines whether his patient satisfies explicit criteria. For example, for a diagnosis of the depressive form of primary affective disorder the three requirements are dysphoric mood, a psychiatric illness lasting at least one month with no other pre-existing psychiatric condition,and at least five of the following eight symptoms: poor appetite or weight loss; sleep difficulty; loss of energy; agitation or retardation; loss of interest in usual activities or decrease in sexual drive; feelings of self-reproach or guilt; complaints of or actually diminished ability to think or concentrate; and thoughts of death or suicide.

A consequence of the St. Louis approach is the necessity for an ‘undiagnosed psychiatric disorder’ category for those patients who do not meet any of the criteria for the specified diagnoses. In actual use, this category is applied to 20-30 per cent of newly-admitted in-patients. These two approaches, structuring the interview and specifying all diagnostic criteria, are being merged in a series of collaborative studies on the psychobiology of the depressive disorders sponsored by the N.I.M.H. Clinical Research Branch. We are confident that this merging will result not only in improved reliability but in improved validity which is, after all our ultimate goal.

And so the DSM-III and all that followed was created: kappa, the Feighner criteria, structured interviews.

In Baseball, the uniforms change, the stadiums grow, the ticket and hot dog prices escalate with the players’ salaries, and the records very slowly fall. But the box scores stay the same and the distance from home to first base never changes. When you have a losing season, you don’t argue with the Stats, you get a new Manager, add a relief pitcher, but you sure don’t fight with the way the game is played.

The DSM-5 tried to change the game with a ‘paradigm shift’ to clinical neuroscience as the sole organizer of psychiatry, and it just didn’t fly. Now they’re telling us that there’s some reason to redefine the meaning or interpretation of kappa. Foul Ball! So I colored in Dr. Frances’ table according to Spitzer, circa 1974. Here are the box scores:
  1.  
    Ivan
    May 8, 2012 | 2:42 AM
     

    There is also some shifting of the goalposts happening this week at the APA meeting in Philadelphia. The DSM-5 leaders have announced a new ordinal scale of Kappa values that “spins” their field trial results for maximum public relations effect. As you said in your final Table in self evident, lousy Kappas are now being deemed ‘acceptable’ and poor Kappas are being called ‘realistic.’

    It’s a bit like Humpty Dumpty: ‘When I use a Kappa,’ Humpty Dumpty said, in rather a scornful tone, ‘it means just what I choose it to mean — neither more nor less.’ ‘The question is,’ said Alice, ‘whether you can make Kappas mean so many different things.’

  2.  
    Tom
    May 8, 2012 | 6:49 AM
     

    You know in psychometric assessment, kappa’s below 0.80 are unacceptable.

  3.  
    May 8, 2012 | 6:55 AM
     

    I didn’t know the actual cut-off, but I knew it was much higher than these people were talking about. I wish the psychometrist Gurus would weigh in here. I think it would help with some much needed reality testing…

  4.  
    Tom
    May 8, 2012 | 8:09 PM
     

    OK psychometric testing is a high standard. But the state of the art as to reliability (rater agreement ) is offered as:

    Interpreting the Magnitude of Kappa
    Landis and Koch (45) have proposed the following as standards for strength of agreement for the kappa coefficient: ≤0=poor, .01–.20=slight, .21–.40=fair, .41–.60=moderate, .61–.80=substantial, and .81–1=almost perfect. Similar formulations exist,(46–48) but with slightly different descriptors. The choice of such benchmarks, however, is inevitably arbitrary,(29,49) and the effects of prevalence and bias on kappa must be considered when judging its magnitude.

    Even by this standard, the DSM-5 reliability studies are woeful in terms of results. Or maybe not: If we accept the reliability of psychiatric diagnosis as FAIR being our standard, then the DSM-5 did its job!!!!

  5.  
    May 8, 2012 | 11:14 PM
     

    Londoners were better than New Yorkers at recognizing schizophrenia? !!

  6.  
    May 8, 2012 | 11:18 PM
     

    Interesting. All the big money-makers, pharma-wise, have crappy kappas in DSM-5.

    Overreaching come to roost?

  7.  
    May 8, 2012 | 11:28 PM
     

    News on the DSM5 from NYT

    Psychiatry Manual Drafters Back Down on Diagnoses

    http://www.nytimes.com/2012/05/09/health/dsm-panel-backs-down-on-diagnoses.html?_r=2&smid=tw-nytimes&seid=auto

Sorry, the comment form is closed at this time.