Home A Multiobserver Study of the Effects of Including Point-of-care Patient Photographs with Portable Radiography
Post
Cancel

A Multiobserver Study of the Effects of Including Point-of-care Patient Photographs with Portable Radiography

Rationale and Objectives

To evaluate whether the presence of facial photographs obtained at the point-of-care of portable radiography leads to increased detection of wrong-patient errors.

Materials and Methods

In this institutional review board–approved study, 166 radiograph–photograph combinations were obtained from 30 patients. Consecutive radiographs from the same patients resulted in 83 unique pairs (ie, a new radiograph and prior, comparison radiograph) for interpretation. To simulate wrong-patient errors, mismatched pairs were generated by pairing radiographs from different patients chosen randomly from the sample. Ninety radiologists each interpreted a unique randomly chosen set of 10 radiographic pairs, containing up to 10% mismatches (ie, error pairs). Radiologists were randomly assigned to interpret radiographs with or without photographs. The number of mismatches was identified, and interpretation times were recorded.

Results

Ninety radiologists with 21 ± 10 (mean ± standard deviation) years of experience were recruited to participate in this observer study. With the introduction of photographs, the proportion of errors detected increased from 31% (9 of 29) to 77% (23 of 30; P = .006). The odds ratio for detection of error with photographs to detection without photographs was 7.3 (95% confidence interval: 2.29–23.18). Observer qualifications, training, or practice in cardiothoracic radiology did not influence sensitivity for error detection. There is no significant difference in interpretation time for studies without photographs and those with photographs (60 ± 22 vs. 61 ± 25 seconds; P = .77).

Conclusions

In this observer study, facial photographs obtained simultaneously with portable chest radiographs increased the identification of any wrong-patient errors, without substantial increase in interpretation time. This technique offers a potential means to increase patient safety through correct patient identification.

The Institute of Medicine’s Quality Report estimates that nearly 98,000 deaths annually may be attributed to medical errors . Within radiology, an important source of error is the wrong-patient error , wherein one patient’s imaging study may be placed in another patient’s folder in a picture archiving and communication system (PACS). Such errors can create problems for both the involved patients. In fact, the National Quality Forum recognizes that wrong-patient errors can affect radiologic practice and that agency, along with the Agency for Healthcare Research and Quality, specifically endorses implementation of a “standardized protocol to prevent mislabeling of radiographs” .

Based on a statewide adverse event reporting system, the Pennsylvania Patient Safety Authority published that, in 2009, it received 652 reports on radiology events: 196 (30.1%) of these reported events were related to wrong-patient events. Of these 196 wrong-patient events, 93 (47%) occurred in the imaging modality of radiography. The report concluded that such errors “occur more frequently than health care providers and patients may realize” despite various quality improvement measures .

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Materials and methods

Study Population

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Data Acquisition and Case Selection

Get Radiology Tree app to read full this article<

Figure 1, Hanging protocol displaying a study as seen by some of the observers. (a) The radiograph–photograph combination on the left refers to the present study and shows a 64-year-old white woman with history of aortic stenosis and aortic valve replacement. (b) The radiograph–photograph combination on the right refers to the prior study, which shows a radiograph obtained 2 months earlier, of a 73-year-old man also with a history of aortic stenosis and aortic valve replacement. The radiographic differences are subtle and mainly related to body habitus. The presence of the photographs show more obvious differences in hair color and body habitus, and the observer asked to interpret this case correctly identified the wrong-patient error.

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Display Workstation Environment

Get Radiology Tree app to read full this article<

Figure 2, ReviewStudyComponent —image of a pop-up window used by the observers to evaluate the image pairs. The “ Other Comments ” field was used to obtain free-format responses from observers.

Get Radiology Tree app to read full this article<

Figure 3, PostStudyQuestionnaire —image of a pop-up window with questionnaire for observers who interpreted studies with photographs. This questionnaire was completed at the end of the interpretation session.

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Observer Study Design

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Statistical Analysis

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Results

Observer Characteristics

Get Radiology Tree app to read full this article<

Table 1

Observer Characteristics

Characteristic Value without Photos ( n = 45) Value with Photos ( n = 45) Total ( n = 90) Age (years) ∗ 53 ± 10 (35, 83) 53 ± 9 (37, 82) 53 ± 10 (35–83) Gender † Male 27 29 56 (62) Female 18 16 34 (38) Years postresidency training ∗ 22 ± 10(4, 52) 20 ± 10 (6, 50) 21 ± 10 (4–52) Fellowship training † Yes 38 38 76 (84) No 7 7 14 (16) Fellowship subspecialty † Abdominal 11 5 16 (18) Breast 2 4 6 (7) Cardiothoracic 7 9 16 (18) Interventional 5 5 10 (11) Musculoskeletal 2 5 7 (8) Neuroradiology 2 2 4 (4) Nuclear medicine 2 3 5 (6) Pediatric 4 5 9 (10) Other 3 1 4 (4) Not applicable 7 6 13 (14) Current subspecialty practice † Abdominal 11 6 17 (19) Breast 8 7 15 (17) Cardiothoracic 9 9 18 (20) Interventional 4 5 9 (10) Musculoskeletal 3 7 10 (11) Neuroradiology 2 2 4 (4) Nuclear Medicine 2 2 4 (4) Pediatric 3 6 9 (10) General radiology 3 1 4 (4)

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Qualitative Results Examples

Get Radiology Tree app to read full this article<

Figure 4, Representative wrong-patient error without photographs. (a) An 89-year-old white man with a history of aortic stenosis, status post surgical aortic valve replacement; median sternotomy wires are seen with proper window-level settings. (b) Radiograph obtained 3 weeks earlier, used to serve as a comparison for (a) in a wrong-patient error scheme, shows a 63-year-old woman also with a history of aortic stenosis, status post percutaneous aortic valve replacement. Differences in the prosthetic valves, sternotomy wires, and vascular and nodal calcifications in both patients are evident. It is unlikely that a person who has undergone surgical aortic valve replacement (the patient in b ) would subsequently undergo percutaneous aortic valve replacement and have the medical sternotomy wires removed in the same 3-week period (the patient in a ). Even more illogical is the development of advanced calcified atherosclerotic disease and calcified mediastinal lymph nodes within 3 weeks. Despite these obvious differences, the observer who was shown this combination did not identify the wrong-patient error.

Get Radiology Tree app to read full this article<

Quantitative Results

Get Radiology Tree app to read full this article<

Table 2

Likelihood of Detecting a Wrong-patient Error: All Radiologists

Characteristic Without Photos With Photos Radiologists unexposed to wrong-patient error ( n = 31) 0/16 (0) 0/15 (0) Radiologists exposed to wrong-patient error ( n = 59) 9/29 (31) 23/30 (77)

Data are presented as ratio of errors detected to number of observers in each category and percent in parentheses.

Table 3

Likelihood of Detecting Wrong-patient Error by Radiologist Characteristics

Characteristic Without Photos Errors detected ∗ With Photos Errors Detected Odds Ratio † P Value ‡ Total readers ( n = 59) 9/29 (31) 23/30 (77) 7.30 (2.29–23.18) 0.006 Current subspecialty Cardiothoracic ( n = 12) 1/5 (20) 6/7 (86) 24 (1.14–505.19) 0.072 Noncardiothoracic ( n = 47) 8/24 (33) 17/23 (74) 5.67 (1.61–19.99) 0.008 Portable chest radiograph interpretation Specialties with current expertise in chest radiography ( n = 36) § 6/18 (33) 14/18 (78) 7 (1.5–30.80) 0.017 Specialties without current expertise in chest radiography ( n = 23) ‖ 3/11 (27) 9/12 (75) 8 (1.24–51.50) 0.039 Experience level >10 year postresidency ( n = 50) 8/25 (32) 20/25 (80) 8.5 (2.34–30.98) 0.014 ≤10 year postresidency ( n = 9) 1/4 (25) 3/5 (60) 4.5 (0.25–80.56) 0.524 Fellowship subspecialty Cardiothoracic ( n = 10) 1/3 (33) 7/10 (86) 12 (0.49–294.55) 0.183 Noncardiothoracic ( n = 49) 8/26 (31) 17/23 (74) 6.38 (1.83–22.23) 0.004 Fellowship trained Yes ( n = 51) 8/25 (32) 21/26 (81) 8.93 (2.46–32.33) 0.006 No ( n = 8) 1/4 (25) 2/4 (50) 3 (0.15–59.89) 1.000

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Poststudy Questionnaire

Get Radiology Tree app to read full this article<

Table 4

Results of Questionnaire That Observers Who Interpreted Studies with Photographs Completed at the End of Their Interpretation Session ∗

Question Response Yes No 1) Were the photographs a distraction? 8 (20) 31 (80) 2) Did you feel you spent more time because of the photographs? 18 (42) 25 (58) 3) Did the photographs help with the interpretation? 19 (44) 24 (56) 4) If you noted mismatched photographs, did you go back and check the radiographs? † 22 (96) 1 (4)

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Discussion

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Relationship to Prior Studies

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Study Limitations

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

Conclusions

Get Radiology Tree app to read full this article<

Acknowledgments

Get Radiology Tree app to read full this article<

Get Radiology Tree app to read full this article<

References

  • 1. Committee on Quality of Health Care in America, Institute of Medicine: To err is human: building a safer health system.1999.National Academy PressWashington, DC

  • 2. National Quality Forum. 30 Safe practices for better health care fact sheet. Available at: http://www.ahrq.gov/qual/30safe.htm . Accessed June 23, 2012.

  • 3. Pennsylvania Patient Safety Advisory: Applying the universal protocol to improve patient safety in radiology. Pennsylvania Patient Safety Authority 2011; 8: pp. 63-70.

  • 4. Joint Commission on Accreditation of Healthcare Organizations 2012 National patient safety goals. Available at: http://www.jointcommission.org/assets/1/6/HAP_NPSG_Chapter_2014.pdf . Accessed May 26, 2014.

  • 5. Ramamurthy S.P., Bhatti P., Arepalli C.D., et. al.: Integrating patient digital photographs with medical imaging examinations. Journal of Digital Imaging 2013; 26: pp. 875-885.

  • 6. Tridandapani S, Provenzale J, Salama M, et al. Systems, methods and computer readable storage media storing instructions for generating an image series. U.S. Patent Application No. 2013/0177,222. July 11, 2013.

  • 7. Tridandapani S., Ramamurthy S., Galgano S., et. al.: Increasing rate of detection of wrong patient radiographs: use of photographs obtained at the time of radiography. AJR Am J Roentgenol 2013; 200: pp. W345-W352.

  • 8. Morishita J., Watanabe H., Katsuragawa S., et. al.: Investigation of misfiled cases in the PACS environment and a solution to prevent filing errors for chest radiographs. Acad Radiol 2005; 12: pp. 97-103.

  • 9. Schultz S., Watson R., Prescott S., et. al.: Patient safety event reporting in a large radiology department. AJR Am J of Roentgenol 2011; 197: pp. 684-688.

  • 10. Versel, N. Making the most of patient safety I.T. Health Data Management Magazine. Available at: http://www.healthdatamanagement.com/issues/19-8/making-the-most-of-patient-safety-i.t.-42891-1.html . Accessed February 25, 2014.Pubmed Partial Author STitle

  • 11. Horowitz, B. Fingerprint biometrics allow hospitals to securely access e-prescriptions, EHRs. eWeek 2011. Available at: http://www.eweek.com/c/a/Health-Care-IT/Fingerprint-Biometrics-Allow-Hospitals-to-Securely-Access-ePrescriptions-EHRs-386667 . Accessed February 25, 2014.

  • 12. Anderson H. Iris scans for patient ID. Healthcare Infosecurity 2010. Available at: http://www.healthcareinfosecurity.com/articles.php?art_id=2439 . Accessed February 25, 2014.

  • 13. Brown C.L.: Health-care data protection and biometric authentication policies: comparative culture and technology acceptance in China and in the United States. Review of Policy Research 2012; 29: pp. 141-159.

  • 14. Morishita J., Katsuragawa S., Konda K., et. al.: An automated patient recognition method based on an imaging matching technique using previous chest radiographs in the picture archiving and communication environment. Med Physics 2001; 28: pp. 1093-1097.

  • 15. Kao E.-F., Lin W.-C., Jaw T.-S., et. al.: Automated patient identify recognition by analysis of chest radiograph features. Acad Radiol 2013; 20: pp. 1024-1031.

  • 16. Turner Y., Hadas-Halpern I.: “The effects of including a patient’s photograph to the radiographic examination.”.Radiological Society of North America scientific assembly and annual meeting.2008.Radiological Society of North AmericaOak Brook, Ill:pp. 576.

  • 17. Weiss F., Safdar N.: Do digital pictures add to the interpretation of diagnostic imaging: a survey along the spectrum of experience. AJR Am J Roentgenol 2010; 194: pp. A5.

  • 18. Slasky B.S., Gur D., Good W.F., et. al.: Receiver operating characteristic analysis of chest image interpretation with conventional, laser-printed, and high-resolution workstation images. Radiology 1990; 174: pp. 775-780.

This post is licensed under CC BY 4.0 by the author.