Impact of Missing Data in Training Artificial Neural Networks for Computer-Aided Diagnosis Mia K. Markey, Ph.D. Biomedical Engineering The University of Texas at Austin [email protected]

Amit Patel Electrical and Computer Engineering The University of Texas at Austin [email protected]

Abstract

This approach is most compelling when the Breast Imaging and Reporting Data System (BIRADS™) lexicon [11] is used since in principle these data are routinely collected already. However, one limitation of the work to date in this area is that carefully curated databases with minimal missing values have been collected for research purposes, yet most clinical databases in routine practice are missing some pieces of information. The purpose of this study was to explore the impact of missing data in the training of artificial neural network (ANN) models for the task of predicting whether breast lesions are benign or malignant based on their BI-RADS™ descriptors. We investigated the form of ANNs that has been most commonly applied in prior breast cancer CAD studies, the feed-forward, back-propagation ANN (BP-ANN). In order to use the BP-ANN, estimates are needed for missing values. Two methods for estimating the missing values were compared: simply replacing missing values with zero and replacing missing values with the mean value from the training set. We compared our results to those obtained in our previous study of the impact of missing data in evaluating ANN models for CAD when the models were trained on complete data sets (manuscript in review).

Artificial neural networks (ANN) are frequently used in the development of Computer-Aided Diagnosis systems for breast cancer detection and diagnosis. One class of models uses descriptions of mammographic lesions encoded following the BI-RADS™ lexicon. Data sets that have been carefully curated to ensure completeness are generally used; however, in routine practice, some information is typically missing in clinical databases. The impact of missing data on the performance of a feedforward, back-propagation ANN, as measured by the area under the Receiver Operating Characteristic curve, was found to be much higher when data were missing from the testing set than when data were missing from the training set. This empirical study highlights the need for additional research on developing robust clinical decision support systems for realistic environments in which key information may be unknown or inaccessible.

1.

Introduction

Among American women, breast cancer is the most common cancer, excluding skin cancers, and is the second leading cause of cancer deaths, after lung cancer [1]. Early detection via mammography improves survival [2]. However, mammography is not perfect and improvements in both the sensitivity and specificity of the exam are needed. One approach is to develop computer-aided detection and diagnosis (CAD) systems to aid radiologists in the interpretation of mammograms [3-5]. Some previous studies in breast cancer CAD have explored the use of statistical and machine learning models for predicting the pathology of breast lesions from experts’ descriptions of mammographic findings [6-10].

2.

Materials and Methods

1.

Data Set

The data set consisted of 604 nonpalpable, mammographically suspicious breast masses that underwent biopsy (core or excisional) at Duke University Medical Center from 1990 to 2000. The pathology outcome was coded as a binary variable of benign vs. malignant. Experienced mammographers described each case using the Breast Imaging

and Reporting Data System (BI-RADS™) lexicon [11]. In particular, the descriptors used in this study were mass margin, mass shape, mass size, and mass density, which were numerically encoded as described in our previous studies [12]. Patient age was also included since it was found to be an important variable in prior analyses of CAD systems that utilize BI-RADS™ descriptors [13]. The data were randomly partitioned into two sets of equal size, A and B. Only cases for which all variable values were present were included in this study.

2.

Missing Data Estimation

Alternate versions of each data set (A, B) were created in which a fraction of the variable values were missing. The amount of missing data as a fraction of the total number of variable values (302 cases x 5 = 1510) was 10%, 20%, 30%, or 40%. In the remainder of the paper, the notation “A0” will be used to indicate set A with 0% missing, “A10” to indicate set A with 10% missing, and so on. Two methods for estimating the missing values were compared: simply replacing missing values with zero and replacing missing values with the mean value.

3.

! !

BP-ANN

A feed-forward back-propagation artificial neural network (BP-ANN) was trained to predict the biopsy outcome from the BIRADS™ descriptors and patient age. BP-ANN is common machine learning algorithm that has been described in several excellent textbooks (e.g., [14, 15]). Briefly, the output of each neuron in a BP-ANN is the result of an !x activation function ( y = 1/(1 + e ) ) applied to a weighted sum of the inputs to the neuron. The weights are the parameters adjusted as the network learns a given task. The ANN is feedforward in the sense that each neuron in one layer feeds into each neuron in the next layer. The BP-ANN is trained to minimize the mean of the sum-of-squares error ( MSE ) using the back-propagation algorithm. The MSE is the squared difference between the network output ( y i " (0,1) ) and network target ( t i " {0,1} ), averaged over all!of the cases ( N , indexed by i ). The back-propagation ! algorithm details how the error should be propagated back through the network to adjust the weights. !

!

The number of hidden nodes, learning rate, momentum, and number of iterations were empirically optimized through leave-one-out training on A0. The network parameters determined were then used for the remainder of the study. Custom software in the C language was used to implement the BP-ANN.

4.

ROC

Receiver Operating Characteristic (ROC) curves can be used to show the trade-off in sensitivity and specificity achievable by a classifier by varying the threshold on the output decision variable [16, 17]. Sensitivity or the true positive fraction (TPF) is the fraction of positive cases that were classified correctly as positive. The specificity, or one minus the false positive fraction (FPF), is the fraction of negative cases that were correctly classified as negative. An ROC curve is generated by applying a threshold to the output of a classification scheme and then plotting the (FPF, TPF) pairs for each threshold. The performance of classification methods can be compared in terms of indices calculated from their curves. In particular, the area under the ROC curve (AUC) is often used as a measure of classifier performance. The values for AUC range from 0.5 for chance to 1.0 for a perfect classifier. In this study, empirical (nonparametric) ROC curves were used and the AUC was numerically integrated using the trapezoid rule. The standard errors and paired statistical comparisons were computed using bootstrap sampling on the classifier outputs. Custom software in the C language was used to implement the ROC analyses.

5.

Summary of Experimental Design

Following either zero imputation or mean imputation to estimate the missing values in the training set, a BP-ANN model was trained on A0 and tested on B0, trained on A10 and tested on B0, and so forth. The performance of the models trained using the zero imputation method was also compared on sets B10, B20, B30, and B40.

3.

Results and Discussion

The purpose of this study was to explore the impact of missing data in the training of artificial neural network models for the task of predicting whether breast lesions are benign or malignant based on their BI-RADS™

descriptors. Two methods for estimating the missing values were compared: simply replacing missing values with zero and replacing missing values with the mean value. The learning parameters and network structure for a BP-ANN model were determined empirically by leave-one-out training on set A with no missing values, i.e., A0, for the task of predicting biopsy outcome of breast masses from BI-RADS™ descriptors and patient age. The selected parameters were then used to train the BP-ANN on set A0. Likewise, the same network parameters were then used to train the BP-ANN on sets A10, A20, A30, and A40 following estimation of the missing values by either simply replacing the missing values with zeros or replacing the missing values with the mean. The networks were then tested on set B with no missing values, i.e., B0. In addition, the networks were tested on data set B with various levels of missing data (B10, B20, B30, B40). The missing data in the test sets were estimated in the same way as on the training sets. The AUC under ROC curve for all the analyses is summarized in Table 1. The BP-ANN models were found to be remarkably robust to the presence of missing data in the training set, even when the missing values were simply replaced by zeros. No benefit was seen to replacing the missing values with the mean, rather than zero, for the levels of missing data studied. Moreover, the BP-ANN’s trained in the presence of missing data were found to be as robust to missing data in the testing as was a BPANN trained on complete data. The relatively minor effects of missing values in the training set are in sharp contrast to our prior study that demonstrated the dramatic impact that missing data can have in the performance of an ANN that was trained on a complete data set (manuscript in review). It is encouraging that even very simple imputation methods may be adequate for some applications, such as was shown here. One implication of this study is that more attention may need to be given to the environment in which a CAD model will actually be used, while less care may be needed than is currently devoted to the curation of data sets for model creation. This empirical study highlights the need for additional research on developing robust clinical decision support systems for realistic environments in which key information may be unknown or inaccessible.

Table 1. Performance in terms of ROC AUC for ANN models trained and tested on data sets with various percentages of missing values. The missing values were estimated two ways: simply replacing missing values with zero (“0”) or replacing missing values with the mean value from the training set (“Ave”). The same method for replacing missing values was used on both the training and testing sets. Train

Test

BP-ANN (0)

BP-ANN (Ave) 0.94 ± 0.01

A0 A0 A0 A0 A0

B0 B10 B20 B30 B40

0.94 ± 0.01 0.84 ± 0.03 0.78 ± 0.03 0.69 ± 0.03 0.72 ± 0.03

A10 A10 A10 A10 A10

B0 B10 B20 B30 B40

0.93 ± 0.02 0.87 ± 0.02 0.81 ± 0.03 0.72 ± 0.03 0.74 ± 0.03

0.95 ± 0.01

A20 A20 A20 A20 A20

B0 B10 B20 B30 B40

0.94 ± 0.02 0.87 ± 0.02 0.82 ± 0.03 0.70 ± 0.04 0.73 ± 0.03

0.94 ± 0.01

A30 A30 A30 A30 A30

B0 B10 B20 B30 B40

0.94 ± 0.02 0.87 ± 0.03 0.82 ± 0.03 0.69 ± 0.04 0.71 ± 0.03

0.94 ± 0.01

A40 A40 A40 A40 A40

B0 B10 B20 B30 B40

0.93 ± 0.02 0.87 ± 0.02 0.82 ± 0.03 0.71 ± 0.03 0.75 ± 0.03

0.94 ± 0.01

4.

Acknowledgements

The authors would like to thank Al Daniel for scientific programming assistance.

5.

References

1.

American Cancer Society, Cancer Facts and Figures 2004. 2004: Atlanta. C.H. Lee, "Screening mammography: proven benefit, continued controversy", Radiologic clinics of North America, 2002, 40: p. 395-407.

2.

3. 4.

5.

6.

7.

8.

9.

10.

11.

12.

13.

14. 15. 16. 17.

M.L. Giger, "Computer-aided diagnosis of breast lesions in medical images", Computing in Science & Engineering, 2000, 2(5): p. 39-45. M.L. Giger, N. Karssemeijer, and S.G. Armato, III, "Computer-aided diagnosis in medical imaging", IEEE Transactions on Medical Imaging, 2001, 20(12): p. 1205-1208. C.J. Vyborny, M.L. Giger, and R.M. Nishikawa, "Computer-aided detection and diagnosis of breast cancer", Radiologic Clinics of North America, 2000, 38(4): p. 725-40. Y. Wu, M.L. Giger, K. Doi, C.J. Vyborny, R.A. Schmidt, and C.E. Metz, "Artificial neural networks in mammography: application to decision making in the diagnosis of breast cancer", Radiology, 1993, 187(1): p. 81-7. C.E. Kahn, Jr., L.M. Roberts, K.A. Shaffer, and P. Haddawy, "Construction of a Bayesian network for mammographic diagnosis of breast cancer", Computers in Biology & Medicine, 1997, 27(1): p. 19-29. C.E. Floyd, Jr., J.Y. Lo, and G.D. Tourassi, "Case-based reasoning computer algorithm that uses mammographic findings for breast biopsy decisions", American Journal of Roentgenology, 2000, 175(5): p. 1347-52. J.A. Baker, P.J. Kornguth, J.Y. Lo, and C.E. Floyd, Jr., "Artificial neural network: improving the quality of breast biopsy recommendations", Radiology, 1996, 198(1): p. 131-5. G.D. Tourassi, M.K. Markey, J.Y. Lo, and C.E. Floyd, Jr., "A neural network approach to breast cancer diagnosis as a constraint satisfaction problem", Medical Physics, 2001, 28(5): p. 80411. American College of Radiology, Illustrated breast imaging reporting and data system (BI-RADSTM). Third ed. 1998, Reston, VA: American College of Radiology. M.K. Markey, J.Y. Lo, G.D. Tourassi, and C.E. Floyd, Jr., "Self-organizing map for cluster analysis of a breast cancer database", Artificial Intelligence in Medicine., 2003, 27(2): p. 113-27. J.Y. Lo, J.A. Baker, P.J. Kornguth, and C.E. Floyd, Jr., "Effect of patient history data on the prediction of breast cancer from mammographic findings with artificial neural networks", Academic Radiology., 1999, 6(1): p. 10-5. T.M. Mitchell, Machine Learning. McGraw-Hill Series in Computer Science, ed. C.L. Liu. 1997, Boston: WCB/McGraw-Hill. 414. C.M. Bishop, Neural Networks for Pattern Recognition. 1995, New York: Oxford University Press. C.E. Metz, "Basic principles of ROC analysis", Seminars in Nuclear Medicine, 1978, 8(4): p. 28398. C.E. Metz, "ROC methodology in radiologic imaging", Investigative Radiology., 1986, 21(9): p. 720-33.

Impact of Missing Data in Training Artificial Neural ...

by the area under the Receiver Operating .... Custom software in the C language was used to implement the BP-ANN. 4. ROC. Receiver Operating Characteristic.

226KB Sizes 1 Downloads 233 Views

Recommend Documents

Impact of missing data in evaluating artificial neural networks trained ...
Feb 17, 2005 - mammographic Breast Imaging and Reporting Data System ... Keywords: Diagnosis; Computer-assisted; Mammography; Breast neoplasms .... The Lypaponov energy function was used as a measure of the network stability. ..... Electrical and Ele

Review Paper on Artificial Neural Network in Data ...
networks have high acceptance ability for high accuracy and noisy data and are preferable ... applications such as identify fraud detection in tax and credit card.

pdf-0946\artificial-neural-networks-in-cancer-diagnosis-prognosis ...
... apps below to open or edit this item. pdf-0946\artificial-neural-networks-in-cancer-diagnosi ... gement-biomedical-engineering-from-brand-crc-press.pdf.

Electromagnetic field identification using artificial neural ... - CiteSeerX
resistive load was used, as the IEC defines. This resistive load (Pellegrini target MD 101) was designed to measure discharge currents by ESD events on the ...

ARTIFICIAL NEURAL NETWORK MODELLING OF THE ...
induction furnace and the data on recovery of various alloying elements was obtained for charge ..... tensile tests of the lab melt ingot has been employed.

Prediction of Software Defects Based on Artificial Neural ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 2, Issue .... Software quality is the degree to which software possesses attributes like ...

Artificial Speciation of Neural Network Ensembles
Problems taken from UCI machine learning benchmark repository. – Wisconsin breast cancer dataset. • 699 instances. • 2 classes - malignant or benign.

Development of an artificial neural network software ...
a distance of 1.5m, using a broadband, TEM horn antenna. ... London EC1V 0HB, U.K. .... 6 Comparison between the peaks of the three perpendicular directions.

Artificial Speciation of Neural Network Ensembles
... Khare & Xin Yao. School Of Computer Science ... Datasets divided into - training (1/2th), validation ... Preserving best individual with shared fitness. • Genetic ...

[PDF Online] Fundamentals of Artificial Neural Networks
Online PDF Fundamentals of Artificial Neural Networks (MIT Press), Read PDF .... over 200 end-of-chapter analytical and computer-based problems that will aid ...

pdf-175\artificial-neural-nets-and-genetic-algorithms-proceedings-of ...
... apps below to open or edit this item. pdf-175\artificial-neural-nets-and-genetic-algorithms-p ... nal-conference-in-ales-france-1995-by-david-w-pears.pdf.

Prediction of Software Defects Based on Artificial Neural ... - IJRIT
studied the neural network based software defect prediction model. ... Neural Networks models have significant advantage over analytical models because they ...

Impact of Artificial _Gummy_ Fingers on Fingerprint Systems_2002.pdf
(2) An unregistered finger (an imposter's finger): An attack against authentication systems by an imposter with his own. biometrics is referred to as non-effort forgery.26 Commonly, accuracy of authentication of fingerprint systems are. evaluated by

Efficient Estimation of Quantiles in Missing Data ... - Research at Google
Dec 21, 2015 - n-consistent inference and reducing the power for testing ... As an alternative to estimation of the effect on the mean, in this document we present ... through a parametric model that can be estimated from external data sources.