Rationale and Objectives
The aim of this study was to assess (1) automated analysis methods versus manual evaluation by human experts of three-dimensional proton magnetic resonance spectroscopic imaging (MRSI) data from patients with prostate cancer and (2) the contribution of spatial information to decision making.
Materials and Methods
Three-dimensional proton MRSI was applied at 1.5 T. MRSI data from 10 patients with histologically proven prostate adenocarcinoma, scheduled either for prostatectomy or intensity-modulated radiation therapy, were evaluated. First, two readers manually labeled spectra using spatial information to identify the localization of spectra and neighborhood information, establishing the reference set of this study. Then, spectra were labeled again manually in a blinded and randomized manner and evaluated automatically using software that applied spectral line fitting as well as pattern recognition routines. Statistical analysis of the results of the different approaches was performed.
Results
Altogether, 1018 spectra were evaluable by all methods. Numbers of evaluable spectra differed significantly depending on patient and evaluation method. Compared to automated analysis, the readers made rather binary decisions, using information from neighboring spectra in ambiguous cases, when evaluating MRSI data as a whole. Differences between anatomically blinded and unblinded evaluation were larger than differences between evaluations using blinded data and automated techniques.
Conclusions
An automated approach, which evaluates each spectrum individually, can be as good as an anatomy-blinded human reader. Spatial information is routinely used by human experts to support their final decisions. Automated procedures that consider anatomic information for spectral evaluation will enhance the diagnostic impact of MRSI of the human prostate.
The large-scale measurement of serum prostate-specific antigen in recent years has resulted in the detection of an immense number of prostate carcinomas . In particular, when initial biopsy results are negative, magnetic resonance (MR) imaging (MRI) is applied to visualize the zonal anatomy of the prostate and localize a possible tumor . High-resolution T2-weighted (T2w) MRI performed with pelvic array coils provides good specificity (up to 90%) but low sensitivity (27%–61%) for tumor detection and localization . The use of an endorectal coil for signal reception raises the sensitivity for tumors >1 cm. However, the reported sensitivity in the literature ranges from 27% to 100%, and 32% to 99% for specificity, depending on the size of the examined tumors . Moreover, false-positive results of T2w MRI remain a problem. They are often caused by local signal reduction due to postbiopsy hemorrhage, prostatitis, or previous treatment .
These limitations fostered the inclusion of functional imaging techniques such as diffusion-weighted MRI, dynamic contrast-enhanced MRI, and proton MR spectroscopic imaging ( 1 H MRSI) in diagnostic imaging protocols. Using MRSI, prostate cancer is characterized by increases in cholines (ie, free choline and choline-containing compounds; Cho) and a decrease in citrate levels . Single-center studies have shown that with MRSI supplementing T2w MRI, prostate cancer can be better differentiated from normal glandular tissue than with conventional MRI alone . Recently published results have demonstrated that most tumors in the prostate were missed because of their small sizes . Thus, MRSI with higher spatial resolution is needed. However, the number of obtained spectra will increase. The large set of spectra resulting from a single examination and the demand for extensive postprocessing and expertise required to interpret the data have hampered the broader application of MRSI.
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Methods
Data
Patients
Get Radiology Tree app to read full this article<
Spectroscopic imaging
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Postprocessing of MRSI data
Get Radiology Tree app to read full this article<
Evaluation Procedures
Get Radiology Tree app to read full this article<
Step 1: Visual evaluation of MR spectroscopic and anatomic data (reference data)
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Step 2: Visual evaluation of randomized spectra (blinded reference)
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Step 3: Automated evaluation using spectral fits
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Step 4: Automated evaluation using pattern recognition
Get Radiology Tree app to read full this article<
Statistical Evaluation
Get Radiology Tree app to read full this article<
Visualization
Get Radiology Tree app to read full this article<
Table 1
Comparison of Class Predictions between Readers and Fitting Methods
Comparison Class 1 Class 2 Class 3 Class 4 Class 5e1 vs e2 Class 1 315 86 15 0 0 Class 2 28 317 7 2 0 Class 3 1 11 83 0 1 Class 4 0 2 13 44 6 Class 5 0 0 0 7 80f1 vs f2 Class 1 392 3 7 0 0 Class 2 9 228 9 9 0 Class 3 1 8 144 7 0 Class 4 3 5 21 109 1 Class 5 0 0 2 16 44
e1 , reader 1; e2 , reader 2; f1 , fitting method 1; f2 , fitting method 2.
Comparison of class predictions by two “blinded” readers evaluating the spectra in a single-voxel fashion shows high agreement. Few evaluations differed by more than one class, and overall variation was less than the differences between algorithms of the fitting metabolite signal templates.
Table 2
Similarity of Performance of the Different Processing Methods, as Measured by Kendall’s τ
Method_an__pr__e1_e2__ea__f1__f2__fl Expert anatomic ( an ) 1 0.73 (0.11) 0.72 (0.09) 0.62 (0.11) 0.67 (0.10) 0.73 (0.06) 0.68 (0.06) 0.58 (0.06) Pattern recognition ( pr ) — 1 0.83 (0.06) 0.68 (0.08) 0.77 (0.07) 0.81 (0.04) 0.75 (0.05) 0.64 (0.05) Expert 1 ( e1 ) — — 1 0.84 (0.05) 0.93 (0.03) 0.74 (0.06) 0.68 (0.04) 0.59 (0.06) Expert 2 ( e2 ) — — — 1 0.93 (0.03) 0.63 (0.08) 0.58 (0.06) 0.51 (0.07) Expert average ( ea ) — — — — 1 0.69 (0.07) 0.64 (0.06) 0.54 (0.05) Fitting metabolite 1 ( f1 ) — — — — — 1 0.95 (0.01) 0.81 (0.04) Fitting metabolite 2 ( f2 ) — — — — — — 1 0.85 (0.04) Fitting lines ( fl ) — — — — — — — 1
Values in parentheses are the standard deviations from bootstrapping. The first line, for example, shows that the expert anatomic method had the highest correlations with pr and f1 , both with τ = 0.73. Here, a τ value 1 indicates perfect correlation and a τ value of 0 complete randomness between two methods. Data are visualized in Figure 6 .
Get Radiology Tree app to read full this article<
Results
General
Evaluation times
Get Radiology Tree app to read full this article<
Evaluable data
Get Radiology Tree app to read full this article<
Table 3
Numbers of Spectra Deemed Evaluable in the Different Approaches and Overlap between the Different Evaluation Methods
Method_an__pr__e1_e2__ea__f1__f2 Expert anatomic ( an ) 4516 (100%) 2093 (46.3%) 2108 (46.7%) 1786 (39.6%) 2259 (50.0%) 2306 (51.1%) 2014 (44.6%) Pattern recognition ( pr ) 2093 (84.0%) 2493 (100%) 1897 (76.1%) 1589 (63.7%) 1785 (71.6%) 1785 (71.6%) 1633 (65.5%) Blinded expert 1 ( e1 ) 2108 (78.8%) 1897 (71.0%) 2674 (100%) 2252 (84.2%) 1897 (70.9%) 1906 (71.3%) 1906 (64.1%) Blinded expert 2 ( e2 ) 1786 (77.5%) 1589 (68.9%) 2252 (97.7%) 2305 (100%) 1588 (68.9%) 1599 (69.4%) 1432 (62.1%) Metabolite spectral fitting (jMRUI) 1 ( f1 ) 2259 (47.3%) 1785 (37.3%) 1897 (39.7%) 1588 (33.2%) 4777 (100%) 4723 (98.9%) 4007 (83.9%) Metabolite spectral fitting (CSItools) 2 ( f2 ) 2306 (47.1%) 1785 (36.4%) 1906 (38.9%) 1599 (33.6%) 4777 (96.4%) 4900 (100%) 4010 (81.8%) Line functions spectral fitting (AMARES) ( fl ) 2014 (49.7%) 1633 (40.3%) 1715 (42.3%) 1432 (35.3%) 4007 (98.8%) 4010 (98.9%) 4055 (100%)
Percentages (in parentheses) indicate the amount of overlap between the methods in the respective row. As an example, among the 4516 spectra evaluated in the anatomic inspection of the data ( an , first row), a subset of 44.6% (2014 spectra) could be evaluated by fitting resonance line models ( fl ). Expert 1 and expert 2 labeled 2674 and 2305 spectra, respectively, and 2252 spectra could be labeled by both.
Get Radiology Tree app to read full this article<
Automated Versus Rater
Get Radiology Tree app to read full this article<
Experts’ single-voxel consensus
Get Radiology Tree app to read full this article<
Line fitting
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Pattern recognition
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Single-Voxel Versus Spatial Analysis
Get Radiology Tree app to read full this article<
Experts’ labels
Get Radiology Tree app to read full this article<
Experts’ labels versus automated methods
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Discussion
General
Get Radiology Tree app to read full this article<
Evaluation time
Get Radiology Tree app to read full this article<
Evaluable data
Get Radiology Tree app to read full this article<
Limitations
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Performance of Automated Methods
Get Radiology Tree app to read full this article<
Evaluation of single-voxel spectra
Get Radiology Tree app to read full this article<
Visual inspection
Get Radiology Tree app to read full this article<
Need for a Spatial Analysis
Get Radiology Tree app to read full this article<
Visual inspection of the MRSI data
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Implications
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
Conclusions
Get Radiology Tree app to read full this article<
Get Radiology Tree app to read full this article<
References
1. Jemal A., Siegel R., Xu J., et. al.: Cancer statistics, 2010. CA Cancer J Clin 2010; 60: pp. 277-300.
2. Kurhanewicz J., Swanson M.G., Nelson S.J., et. al.: Combined magnetic resonance imaging and spectroscopic imaging approach to molecular imaging of prostate cancer. J Magn Reson Imaging 2002; 16: pp. 451-463.
3. Scheidler J., Hricak H., Vigneron D.B., et. al.: Prostate cancer: localization with three-dimensional proton MR spectroscopic imaging—clinicopathologic study. Radiology 1999; 213: pp. 473-480.
4. Hricak H., White S., Vigneron D., et. al.: Carcinoma of the prostate gland: MR imaging with pelvic phased array coil versus integrated endorectal-pelvic phased-array coils. Radiology 1994; 193: pp. 703-709.
5. Roethke M.C., Lichy M.P., Jurgschat L., et. al.: Tumorsize dependent detection rate of endorectal MRI of prostate cancer—a histopathologic correlation with whole-mount sections in 70 patients with prostate cancer. Eur J Radiol 2011; 79: pp. 189-195.
6. Perrotti M., Kaufman R.P., Jennings T.A., et. al.: Endo-rectal coil magnetic resonance imaging in clinically localized prostate cancer: is it accurate?. J Urol 1996; 156: pp. 106-109.
7. Miao H., Fukatsu H., Ishigaki T.: Prostate cancer detection with 3-T MRI: Comparison of diffusion-weighted and T2-weighted imaging. Eur J Radiol 2007; 61: pp. 297-302.
8. Quinn S.F., Franzini D.A., Demlow T.A., et. al.: MR imaging of prostate cancer with an endorectal surface coil technique: correlation with whole-mount specimens. Radiology 1994; 190: pp. 323-327.
9. D’Amico A.V., Schnall M., Whittington R., et. al.: Endorectal coil magnetic resonance imaging identifies locally advanced prostate cancer in select patients with clinically localized disease. Urology 1998; 51: pp. 449-454.
10. Reinsberg S.A., Payne G.S., Riches S.F., et. al.: Combined use of diffusion-weighted MRI and 1H MR spectroscopy to increase accuracy in prostate cancer detection. AJR Am J Roentgenol 2007; 188: pp. 91-98.
11. Schiebler M.L., Yankaskas B.C., Tempany C., et. al.: MR imaging in adenocarcinoma of the prostate: interobserver variation and efficacy for determining stage C disease. AJR Am J Roentgenol 1992; 158: pp. 559-562.
12. Weinreb J.C., Blume J.D., Coakley F.V., et. al.: Prostate cancer: sextant localization at MR imaging and MR spectroscopic imaging before prostatectomy—results of ACRIN prospective multi-institutional clinicopathologic study. Radiology 2009; 251: pp. 122-133.
13. Scheenen T.W., Fütterer J., Weiland E., et. al.: Discriminating cancer from noncancer tissue in the prostate by 3-dimensional proton magnetic resonance spectroscopic imaging: a prospective multicenter validation study. Invest Radiol 2011; 46: pp. 25-33.
14. Kelm B.M., Menze B.H., Zechmann C.M., et. al.: Automated estimation of tumor probability in prostate magnetic resonance spectroscopic imaging: pattern recognition vs quantification. Magn Reson Med 2007; 57: pp. 150-159.
15. Pels P., Ozturk-Isik E., Swanson M.G., et. al.: Quantification of prostate MRSI data by model-based time domain fitting and frequency domain analysis. NMR Biomed 2006; 19: pp. 188-197.
16. Scheenen T.W., Klomp D.W., Röll S.A., et. al.: Fast acquisition-weighted three-dimensional proton MR spectroscopic imaging of the human prostate. Magn Reson Med 2004; 52: pp. 80-88.
17. de Beer R., van den Boogaart A., van Ormondt D., et. al.: Application of time-domain fitting in the quantification of in vivo 1H spectroscopic imaging data sets. NMR Biomed 1992; 5: pp. 171-178.
18. Pijnappel W.W.F., Van den Boogaart A., de Beer R., et. al.: SVD-based quantification of magnetic resonance signals. J Magn Reson 1969; 97: pp. 122-134.
19. Naressi A., Couturier C., Devos J.M., et. al.: Java-based graphical user interface for the MRUI quantitation package. MAGMA 2001; 12: pp. 141-152.
20. Kreis R.: Issues of spectral quality in clinical 1H-magnetic resonance spectroscopy and a gallery of artifacts. NMR Biomed 2004; 17: pp. 361-381.
21. Kelm B.M., Menze B.H., Neff T., et. al.: CLARET: a tool for fully automated evaluation of MRSI with pattern recognition methods.Handels H.Erhardt J.Horsch A. et. al.Bildverarbeitung für die Medizin 2006—Algorithmen, Systeme, Anwendungen.2006.SpringerBerlin, Germany:pp. 51-55.
22. Menze B.H., Kelm B.M., Weber M.A., et. al.: Mimicking the human expert: a pattern recognition approach to score the data quality in MRSI. Magn Reson Med 2008; 59: pp. 1457-1466.
23. Bouix S., Martin-Fernandez M., Ungar L., et. al.: On evaluating brain tissue classifiers without a ground truth. Neuroimage 2007; 36: pp. 1207-1224.
24. Kendall M.: Rank Correlation Methods.1948.Charles Griffin & CompanyLondon 160
25. Borg I., Groenen P.: Modern Multidimensional Scaling: Theory and Applications.2nd ed.2005.SpringerNew York 614
26. Fütterer J.J., Scheenen T.W., Heijmink S.W., et. al.: Standardized threshold approach using three-dimensional proton magnetic resonance spectroscopic imaging in prostate cancer localization of the entire prostate. Invest Radiol 2007; 42: pp. 116-122.
27. Jung J.A., Coakley F.V., Vigneron D.B., et. al.: Prostate depiction at endorectal MR spectroscopic imaging: investigation of a standardized evaluation system. Radiology 2004; 233: pp. 701-708.
28. Cosío F.A.: Automatic initialization of an active shape model of the prostate. Med Image Anal 2008; 12: pp. 469-483.
29. Costa M.J., Delingette H., Novellas S., et. al.: Automatic segmentation of bladder and prostate using coupled 3D deformable models. Med Image Comput Comput Assist Interv 2007; 10: pp. 252-260.
30. Zhu Y., Williams S., Zwiggelaar R.: Computer technology in detection and staging of prostate carcinoma: a review. Med Image Anal 2006; 10: pp. 178-199.
31. Shen D., Lao Z., Zeng J., et. al.: Optimized prostate biopsy via a statistical atlas of cancer spatial distribution. Med Image Anal 2004; 8: pp. 139-150.
32. Laudadio T., Pels P., De Lathauwer L., et. al.: Tissue segmentation and classification of MRSI data using canonical correlation analysis. Magn Reson Med 2005; 54: pp. 1519-1529.
33. Görlitz L., Menze B.H., Weber M.A., et. al.: Semi-supervised tumor detection in magnetic resonance spectroscopic images using discriminative random fields.Hamprecht F.A.Schnörr C.Jähne B.Proceedings of the 29th Symposium of the German Association for Pattern Recognition (DAGM 07), Heidelberg, Germany, Pattern Recognition.2007.SpringerBerlin, Germany:pp. 224-233.