Using the method of judgement analysis to address variations in diagnostic decision making
© Hancock et al; licensee BioMed Central Ltd. 2012
Received: 21 December 2011
Accepted: 13 March 2012
Published: 13 March 2012
Heart failure is not a clear-cut diagnosis but a complex clinical syndrome with consequent diagnostic uncertainty. Judgment analysis is a method to help clinical teams to understand how they make complex decisions. The method of judgment analysis was used to determine the factors that influence clinicians' diagnostic decisions about heart failure.
Three consultants, three middle grade doctors, and two junior doctors each evaluated 45 patient scenarios. The main outcomes were: clinicians' decisions whether or not to make a diagnosis of suspected heart failure; the relative importance of key factors within and between clinician groups in making these decisions, and the acceptability of the scenarios.
The method was able to discriminate between important and unimportant factors in clinicians' diagnostic decisions. Junior and consultant physicians tended to use patient information similarly, although junior doctors placed particular weight on the chest X-Ray. Middle-grade doctors tended to use information differently but their diagnostic decisions agreed with consultants more frequently (k = 0.47) than junior doctors and consultants (k = 0.23), or middle grade and junior grade doctors (k = 0.10).
Judgment analysis is a potentially valuable method to assess influences upon diagnostic decisions, helping clinicians to manage the quality assurance process through evaluation of care and continuing professional development.
Heart failure is not a disease per se, but a complex clinical syndrome . Unlike kidney or lung failure there is no easily measured organ function parameter to determine a diagnosis. Diagnosis and management are based on clinicians' decisions, informed by a number of factors including the patient's medical history, physical examination, electrocardiogram and chest x-ray . With early accurate diagnosis and appropriate treatment prognosis can be substantially improved. However, there is evidence to suggest that diagnosis is missed in up to half of cases [3–5]. Diagnosis and management are complex; symptoms are often non-specific and physical signs are difficult to elicit. An echocardiogram, measuring left ventricular ejection fraction, is used to classify common forms of heart dysfunction although various methods of measurement give different results, and there is no definitive result below which heart failure can be confirmed . Managing uncertainty is central to clinical practice and requires the linking of experience and evidence: this places specialists at an advantage, but junior doctors, nurses and general practitioners often make diagnostic decisions about heart failure.
National Service Frameworks (NSFs) and clinical guidelines are fundamental to clinical governance programmes, aiming to improve standards and reduce unacceptable variations in clinical practice . However, neither can be easily applied to complex decision-making and their recommendations need contextualising to be relevant: no process exists to ensure this routinely occurs and variation in care remains common. The lack of information about the sources of variability in diagnosis and care means that variability may be simplistically viewed as inappropriate, but may appropriately reflect individual patient needs, activity recording variations or alternative but valid traditions of care.
Existing methods of assessing clinicians' decision-making tend to polarise around the process of decision-making (e.g. sociological enquiry) or on its outcome (e.g. audit). Clinical examinations or Objective Structured Clinical Examinations (OSCEs) are popular , but are removed from real life situations . Observations of clinicians in practice aim to address these concerns , but the focus is on the decision rather than on the information that informs it. Similarly, 360° appraisals are used to assess performance  although they also focus on decisions made and are vulnerable to poorly informed or biased reports from colleagues . Clinical decisions often involve evaluating multiple cues in situations of uncertainty, thus, to understand and address unacceptable variations in care, assessment should link the process of decision-making (i.e. the factors that inform a decision) and the decision made.
Judgement analysis links the process (i.e. how clinicians use information to reach a decision) and outcome of clinical decisions [13, 14], The method involves identifying information used in a diagnostic or treatment context. This information is expressed as a number of variables or factors used within a series of scenarios. Judgement analysis offers a means of examining the information used by clinicians in their decision making: it can be used to compare a decision to an agreed standard, such as a protocol or guideline, or to that of another clinician or reference group. Results can be used to direct education and training and to help determine practical limitations of guidelines. In essence, judgment analysis provides a systematic variation of the factors that affect decisions and, thus, provides insight about how doctors use information in their decision making, and how and why they vary. Inappropriate variation in health care is of primary concern but is very poorly researched. While it might seems obvious that, in general, senior clinicians might be better diagnosticians than their junior colleagues, seniority is not a marker of accuracy per se. Furthermore it is not obvious how senior clinicians use information to make better decisions; this information could be used to guide education and training. The method of judgment analysis offers an important way of addressing variation in decision making, but is relatively unused. This paper describes pilot work necessary to develop its use and will be of help to other research teams considering using the technique.
Applying judgment analysis to diagnostic decisions about heart failure
The example used in this paper demonstrates how judgement analysis can be used to assess clinicians' diagnostic decisions about heart failure:
(i) quantifying how a range of factors inform clinicians' diagnostic decisions about heart failure
(ii) measuring any differences in these factors within and between groups of consultant, middle grade and junior doctors
(iii) assessing the acceptability and feasibility of using the scenarios.
Ethical guidance was obtained from County Durham and Tees Valley 1 Research Ethics Committee. Given the nature of the study, ethical approval was deemed by the Chair of this committee not to be required. The research was partially funded by an academic grant from Darlington Memorial Hospital. Study conduct conformed to the principles of The Declaration of Helsinki.
Phase 1. Developing clinical scenarios
The Eight Key Variables Included in the Scenarios and their Levels
2. Pitting oedema
Mild, around ankles
Marked, above knees
3. Ischaemic heart disease
4. Heart rate
5. Lung signs
6. Chest x-ray (lungs)
Upper lobe veins enlarged
7. Heart size
Bundle Branch Block
In order that as much ecological validity as possible was retained, the variables were presented so that the scenarios reflected the format found in practice. Thus, while, for example, a heart rate of 60-100 beats per minute (bpm) was considered normal, < 60 bpm low and > 100 bpm high, a scenario which included 'heart rate: high' might read HR 126 bpm. The following is an example:
A 74-year-old patient presents to A&E with breathlessness. She has no history of IHD. She can lay flat without increased breathlessness. She has no ankle oedema. Her heart rate is 120 bpm; ECG shows T wave changes suggestive of ischaemia. There are crackles at both lung bases; CXR shows pulmonary oedema; heart size is normal. This patient has suspected heart failure: Yes/No (circle one).
The 81 scenarios were discussed with a Consultant Cardiologist to ensure that the format reflected real scenarios and which, if any, should be excluded. 41 scenarios were excluded on the basis that they were not plausible, or that they represented a healthy person. Five scenarios were duplicated in order to determine the consistency of individual responses; these repeated scenarios, called 'hold out' cases, were not included in the analysis. Thus, a total of 45 scenarios were included.
Phase 2. Completion of scenarios by participants
Three cardiology consultants, three middle grade doctors (two Registrars and one Senior House Officer), and two junior doctors (Foundation year 1) were sent the same 45 scenarios in the same order in one document, they were asked to note the time it took to complete the scenarios and evaluate the format. Clinicians were encouraged to make decisions as they would in practice. For each scenario they were asked whether or not they would make a diagnosis of suspected heart failure (where heart failure becomes the 'working diagnosis').
In addition to the scenarios, the doctors were asked to complete a number of other questions. These were: (i) participant demographics; (ii) a list of factors to be rated by participants according to their importance in establishing a diagnosis of heart failure, where 1 = not important, 2 = slightly (or occasionally) important, 3 = moderately (or often) important, and 4 = very (or always) important and the variables were: gender, age, orthopnoea, pitting oedema, history of IHD, heart rate, lung signs, CXR (lungs, heart size, ECG, JVP, Gallop rhythm, out of hours consultation (6 pm-6 am), others (specify); (iii) feedback about the scenarios including: how long it took to complete the scenarios; to what extent the scenarios adequately reflected the information used when making diagnoses with real patient [exactly, very well, quite well, not very well or not at all]; how the scenarios could be improved; any other information that would have been useful in the scenarios; whether or not participants would have preferred to have received the scenarios by email; final comments and suggestions.
Phase 3. Data analysis
The dependent variable within the analyses was the decision whether or not to make a diagnosis of suspected heart failure. Data analysis used the Conjoint procedure within SPSS , which generated a relative utility (importance) score for each variable and variable level . As utility scores share a common metric, it is possible to calculate the relative influence of each variable. Utility scores vary between -1 and 1 with zero denoting no influence and larger magnitudes corresponding to a greater negative or positive contribution. Repeated scenarios were not included in the Conjoint analysis.
Descriptive and inferential statistical analyses were conducted using SPSS version 14. Decision repeatability and consistency within and between clinician groups were analysed using categorical agreement analysis (kappa). The kappa score indicates the level of agreement between or within raters on a scale from zero to one, where the level of agreement is poor (< 0.2), fair (≥ 0.2 to < 0.4), moderate (≥ 0.4 to < 0.6), good (≥ 0.6 to < 0.8) or very good (≥ 0.8).
Clinicians' decisions about whether or not to make a diagnosis of suspected heart failure
Agreement Scores (Kappa) for Repeatability and Consistency of Diagnoses of Suspected Heart Failure
Between Groups: consistency of categorical diagnostic decisions (ordinal levels) for 40 scenarios comparing different level of seniority using weighted Kappa
The relative importance of the variables within and between clinician groups
The Average Importance of the Variables by Clinician Group
Middle Grade Dr Score
Junior Dr Score
Ischaemic heart disease
The acceptability of the scenarios
(i) Time Taken
Time taken to complete the scenarios varied substantially: a mean of 28 min and range of 10 to 45 min. Consultants ranged from 10 to 25 min and junior doctors from 25 to 45 min. All three middle grade doctors reported taking 30 min.
(ii) Overall Format
Consultant cardiologists and middle grade doctors reported that the scenarios reflected clinical situations 'quite well' (n = 5) or 'very well' (n = 1), while junior doctors (n = 2) reported that they did so 'exactly'.
(iii) Preference for Email or Hard Copy of Scenarios
Half of all respondents (n = 6) reported a preference for a hard copy rather than electronic copy of the scenarios. There was variability within and between groups.
Clinical judgment analysis links how clinicians make diagnostic decisions with what decisions are made. This study demonstrated considerable variations in heart failure diagnostic decision-making within and between clinical grades. The method was able to discriminate between influential and uninfluential factors diagnostic decisions both within and between clinician groups. The development, completion and analysis of the scenarios was feasible and acceptable to clinicians.
Without a reference standard answer for each scenario it is not possible to assess the accuracy of diagnoses made, or appropriateness of variations. One possible explanation for variability is that senior clinicians possessed experience and extant knowledge to inform their decisions, junior doctors followed protocols and middle grade doctors employed a combination of the two. Clinical guidelines and protocols seek to standardise practice and to eliminate variability in patient care. However, evidence-based guidelines only provide for 'usual' or 'average' clinical scenarios. Some diagnostic variation may have resulted from clinicians' perceptions about individual patient needs within the scenarios. Notwithstanding the different weight placed on the factors informing diagnostic decisions, consultant and middle grade doctor decision-making was similar. However agreement between junior doctors and more senior grades was only poor to fair - a finding that cannot be disregarded as it may reflect real differences in clinical care.
It is unclear why middle grade doctors made markedly different use of diagnostic information. The method's strength is that it unmasks values used in decision-making that may be incorporated into the training context to explore quality-of-care issues.
This demonstration study is based on a small sample and precludes definitive conclusions. Clinicians were asked to complete the scenarios in the same way as they would in practice. While the scenarios were constructed to reflect real clinical situations faced by them, it is not possible to recreate the same pressures as life in clinical practice. Participants' responses to questions about the acceptability of the scenarios supported their current format.
Relevance to practice
Changing roles within health care, with more junior staff taking on greater responsibility, have been accompanied by increased public and professional scrutiny [18, 19]. This has occurred, in part, because patients, society, and the professions need to be assured that individual clinicians are not only qualified, but consistently provide high quality, safe care for patients. All clinical practise involves uncertainty in diagnosis, prognosis and treatment, and adverse health care events causing physical or psychological injury to patients are surprisingly common . In the UK, adverse events take place in about 10% of NHS admissions and cost £2 billion a year; 400 people die or are seriously injured in adverse events; and, over £500 million was paid out in clinical negligence claims in 2004/2005 . Judgement analysis offers a means of quantifying the factors that influence complex decisions made in clinical practice, thus potentially reducing the likelihood of adverse events when linked to continuing professional development.
Clinical judgement analysis has been applied previously in a number of contexts, supporting our findings of substantial clinical variability when decisions involve complexity [15, 22–26]. The method needs developing beyond measuring variations to demonstrate improved consistency and quality of care. Our findings are the first part of a research programme to develop a targeted education and training tool to promote quality and safety within clinical teams.
Variable clinical decision-making has important implications for diagnosis and management. This is particularly important for heart failure in the hospital setting, since junior and middle grade doctors often make diagnostic decisions. While judgement analysis may help explain and quantify diagnostic variation permitting its discussion as a quality issue within the clinical team. Findings may subsequently positively shape future clinical guidelines, in particular identifying areas of variation and contention. Rather than the imposition of an external clinical governance agenda, the use of this method represents an opportunity for clinical teams to lead the quality assurance process and to differentiate between unacceptable and acceptable variability in care. Judgment analysis usefully captures the determinants of clinicians' diagnostic decisions about heart failure. This pilot study demonstrates the potential for the method to facilitate quality assurance within the clinical team by enabling teams to explore variations, reassess educational support, and make appropriate use of (or modify) guidelines. Further adequately-powered research is required to realize this potential and inform clinical management.
We are grateful to the clinicians who gave their time to participate in this study and who provided valuable feedback.
- McDonagh TA, Dargie HJ: What is Heart Failure?. Managing Heart Failure in Primary Care. Edited by: Dargie HJ, McMurray JJV, Poole-Wilson PA. 1996, London: Blackwell Healthcare Communications Ltd, 1-10.Google Scholar
- The task force on acute heart failure of the European Society of Cardiology: Guidelines on Acute Heart Failure. Eur Hear J. 2005, 26 (4): 384-416.View ArticleGoogle Scholar
- Davies MK, Hobbs FDR, Davis RC, et al.: Prevalence of left ventricular systolic dysfunction and heart failure in the Echocardiographic Heart of England Screening study: A population based study. Lancet. 2001, 358: 439-444. 10.1016/S0140-6736(01)05620-3.PubMedView ArticleGoogle Scholar
- Khand A, Shaw M, Gemmel I, Davis RC, et al.: Do discharge codes underestimate hospitalisation due to heart failure? Validation study of hospital discharge coding for heart failure. Eur J Hear Fail. 2005, 7 (5): 792-797. 10.1016/j.ejheart.2005.04.001.View ArticleGoogle Scholar
- Barents M, van der Horst I, Voors A, Davis RC, et al.: Prevalence and misdiagnosis of chronic heart failure in nursing home residents: the role of B-type natruiretic peptides. Neth Hear J. 2008, 16 (4): 123-128. 10.1007/BF03086130.View ArticleGoogle Scholar
- Dargie HJ: What is heart failure?. British Society for Heart Failure Newsletter. 1998, Spring, 2
- Department of Health: . Draft: Research Governance Framework for Health and Social Care. 2003, London: The Stationary Office, 2Google Scholar
- Major DA: OSCEs - seven years on the bandwagon: The progress of an objective structured clinical evaluation programme. Nurse Edu Today. 2005, 25 (6): 442-454. 10.1016/j.nedt.2005.03.010.View ArticleGoogle Scholar
- Newble D: Techniques for measuring clinical competence: objective structured clinical examinations. Med Edu. 2004, 38: 199-200. 10.1111/j.1365-2923.2004.01755.x.View ArticleGoogle Scholar
- LaDuke S: Competency assessments: a case for the nursing interventions classification and the observation of daily work. J Nurs Adm. 2000, 30: 339-40. 10.1097/00005110-200007000-00001.PubMedView ArticleGoogle Scholar
- Whitehouse A, Walzman M, Wall D: Pilot study of 360° assessment of personal skills to inform record of in training assessments for senior house officers. Hosp Med. 2002, 63 (3): 172-175.PubMedView ArticleGoogle Scholar
- McKinley RK, Fraser RC, Baker R: Model for directly assessing clinical competence and performance in revalidation of clinicians. BMJ. 2001, 322: 712-715. 10.1136/bmj.322.7288.712.PubMedPubMed CentralView ArticleGoogle Scholar
- Heverly MA, Fitt DX, Newmand FL: Constructing case vignettes for evaluating clinical judgments: an empirical model. Eval Programme Plann. 1984, 7: 45-55. 10.1016/0149-7189(84)90024-7.View ArticleGoogle Scholar
- Hammond KR, Stewart TR, Brehmer B, Davis RC, et al.: Social Judgment Theory. Human Judgment and Decision Processes. Edited by: Katz MF, Schwartz S. 1975, New York: Academic, 271-312.View ArticleGoogle Scholar
- Backlund L, Danielsson B, Bring J, Davis RC, et al.: Factors influencing GPs decisions on the treatment of hypercholesterolaemia. Scand J Primary Health Care. 2000, 18: 87-93. 10.1080/028134300750018963.View ArticleGoogle Scholar
- Brunswik E: Perception and the representative design of psychological experiment. 1956, University of California Press: Berkeley, 2Google Scholar
- Statistical Package for Social Scientists (SPSS): SPSS Conjoint 14.0. 2005, USA SPSS IncGoogle Scholar
- Department of Health: Guidance on Working Hours for Junior Doctor. 2002, London: The Stationery OfficeGoogle Scholar
- European Directive: Directive 2003/88/EC of the European Parliament and of the Council of 4th November 2003 concerning certain aspects of the organisation of working time. 2003, Council of the European UnionGoogle Scholar
- Department of Health: An Organisation With A Memory. Report of an expert group on learning from adverse events in the NHS chaired by the Chief Medical Office. 2000, London: The Stationery OfficeGoogle Scholar
- Department of Health: National Health Service Redress Bill. 2005, The Stationery Office, London, [http://www.dh.gov.uk/en/Publicationsandstatistics/Legislation/Regulatoryimpactassessment/DH_4138881]Google Scholar
- Vancheri F, Alletto M, Curcio M: Is clinical diagnosis of heart failure reliable? Clinical judgment of cardiologists versus internists. Eur J Intern Med. 2003, 14: 26-31. 10.1016/S0953-6205(02)00184-X.PubMedView ArticleGoogle Scholar
- Skaner Y, Bring J, Bengt U, Davis RC, et al.: Heart Failure Diagnosis in primary health care: clinical characteristics of problematic patients. A clinical judgment analysis study. BMC Fam Pract. 2003, 4: 12-10.1186/1471-2296-4-12.PubMedPubMed CentralView ArticleGoogle Scholar
- Thompson CA, Foster A, Cole I, Davis RC, et al.: Using social judgment theory to model nurses' use of clinical information in critical care education. Nurse Edu Today. 2005, 25: 68-77. 10.1016/j.nedt.2004.10.003.View ArticleGoogle Scholar
- Harries C, Evans J, Dennis I: Measuring Doctors' Self-insight into their Treatment Decisions. Appl Cogn Psychol. 2000, 14: 455-477. 10.1002/1099-0720(200009)14:5<455::AID-ACP667>3.0.CO;2-V.View ArticleGoogle Scholar
- Kushniruk AW, Patel VL: Cognitive and usability engineering for the evaluation of clinical information systems. J Biomed Inf. 2004, 37: 56-76. 10.1016/j.jbi.2004.01.003.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.