338

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

Robust Learning-Based Parsing and Annotation of Medical Radiographs Yimo Tao, Zhigang Peng, Arun Krishnan, and Xiang Sean Zhou*

Abstract—In this paper, we propose a learning-based algorithm for automatic medical image annotation based on robust aggregation of learned local appearance cues, achieving high accuracy and robustness against severe diseases, imaging artifacts, occlusion, or missing data. The algorithm starts with a number of landmark detectors to collect local appearance cues throughout the image, which are subsequently verified by a group of learned sparse spatial configuration models. In most cases, a decision could already be made at this stage by simply aggregating the verified detections. For the remaining cases, an additional global appearance filtering step is employed to provide complementary information to make the final decision. This approach is evaluated on a large-scale chest radiograph view identification task, demonstrating a very high accuracy ( 99 9%) for a posteroanterior/anteroposterior (PA–AP) and lateral view position identification task, compared with the recently reported large-scale result of only 98.2% (Luo, et al., 2006). Our approach also achieved the best accuracies for a three-class and a multiclass radiograph annotation task, when compared with other state of the art algorithms. Our algorithm was used to enhance advanced image visualization workflows by enabling content-sensitive hanging-protocols and auto-invocation of a computer aided detection algorithm for identified PA–AP chest images. Finally, we show that the same methodology could be utilized for several image parsing applications including anatomy/ organ region of interest prediction and optimized image visualization. Index Terms—Chest radiograph, hanging protocol, medical image annotation, object recognition, picture archive and communication system (PACS).

I. INTRODUCTION HE amount of medical image data produced nowadays is constantly growing, and a fully automatic image content annotation algorithm can significantly improve the image reading workflow, by automatic configuration/optimization of image display protocols, and by offline invocation of image processing (e.g., denoising or organ segmentation) or computer aided detection (CAD) algorithms. However, such annotation

T

Manuscript received June 07, 2010; revised September 01, 2010; accepted September 05, 2010. Date of publication September 27, 2010; date of current version February 02, 2011. This work was done while Y. Tao was a research intern at Siemens Medical Solutions USA. Asterisk indicates corresponding author. Y. Tao is with the Microsoft Health Solutions Group, Chevy Chase, MD 20815 USA (e-mail: [email protected]). Z. Peng and A. Krishnan are with the Siemens Medical Solutions USA, Inc., Malvern, PA 19355 USA (e-mail: [email protected]; arun.krishnan @siemens.com). *X. S. Zhou is with the Siemens Medical Solutions USA, Inc., Malvern, PA 19355 USA (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TMI.2010.2077740

algorithm must perform its tasks in a very accurate and robust manner, because even “occasional” mistakes can shatter users’ confidence in the system, thus reducing its usability in the clinical settings. In the radiographic exam routine, chest radiograph comprises at least one-third of all diagnostic radiographic procedures. Chest radiograph provides sufficient pathological information about cardiac size, pneumonia-shadow, and mass-lesions, with low cost and high reproducibility. However, the projection and orientation information in the DICOM header of radiographic images are often missing or mislabeled in the picture archive and communication system (PACS) [2]. Given a large number of radiographs to review, the accumulated time and cost can be substantial for manually identifying the projection view and correcting the image orientation for each radiograph. The goal of this study is to develop a highly accurate and robust algorithm for automatic annotation of medical radiographs based on the image data, correcting potential errors or missing tags in the DICOM header. Such an algorithm would improve the efficiency and effectiveness of image management, and expedite workflow in hospitals. One specific application of this work is to automatically recognize the projection view of chest radiographs into posteroanterior/anteroposterior (PA–AP) and lateral (LAT) views. Such classification could be exploited on a PACS workstation to support optimized image hanging-protocols [1]. Furthermore, if a chest X-ray CAD algorithm is available, it can be invoked automatically on the appropriate view(s), saving users’ manual effort to invoke such an algorithm and the potential idle time while waiting for the CAD outputs. A. Related Works A great challenge for automatic medical image annotation is the large visual variability across patients in medical images from the same anatomy category. The variability caused by individual body conditions, patient ages, and diseases or artifacts would fail many seemingly plausible heuristics or methods based on global or local image content descriptors. Figs. 1 and 2 show some examples of PA–AP and LAT chest radiographs. Because of obliquity, tilt, differences in projection, and the degree of lung inflation, the same class PA–AP and LAT images may present very high inter patient variability. Fig. 3 shows another example of images from the “pelvis” class with considerable visual variation caused by differences in contrast, field of view (FoV), diseases/implants, and imaging artifacts. Most existing methods (e.g., [3]–[5]) for automatic medical image annotation were based on different types of image content descriptors, separately or combined together with different classifiers. Müller et al. [6] proposed a method using weighted combinations of different global and local features to compute

0278-0062/$26.00 © 2010 IEEE

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

339

Fig. 1. The PA–AP chest images of (a) normal patient, (b) and (c) patients with severe chest disease, and (d) an image with unexposed region on the boundary.

Fig. 2. The LAT chest images of (a) normal patient, (b) and (c) patients with severe chest disease, and (d) an image with body rotation.

the similarity scores between the query image and the reference images in the training database. The annotation strategy was based on the GNU Image Finding Tool image retrieval engine. Güld and Deserno [7] extracted pixel intensities from down-scaled images and other texture features as the image content descriptors. Different distance measures were computed and summed up in a weighted combination form as the final similarity measurement used by the nearest-neighbor decision rule (1NN). Deselaers and Ney [4] used a bag-of-features (BOF) approach based on local image descriptors. The histograms generated using bags of local image features were classified using discriminative classifiers, such as support vector machine (SVM) or 1NN. Keysers et al. [8] used a nonlinear model considering local image deformations to compare images. The deformation measurement was then used to classify the image using 1NN. Tommasi et al. [9] extracted features using the algorithm of scale-invariant feature transform (SIFT) [10] from downscaled images, and then used the similar BOF approach [4]. A modified SVM integrating the BOF and pixel intensity features was used for classification. Avni et al. [5] proposed a method based on local image sub-patches and a BOF approach, with a kernel based SVM classifier. Regarding the task of recognizing the projection view and orientation of chest radiographs, many domain specific algorithms have been proposed. Analysis of image projection profile (e.g., [11]–[13]) is one of the most often applied methods. Pietka and Huang [12] proposed a method using two projection profiles to differentiate PA–AP and LAT chest radiographs. Kao et al. [13] proposed a method based on linear discriminant analysis (LDA) with two features extracted from horizontal axis projection profile. Besides projection profile, many other techniques have also been investigated. Arimura et al. [14] proposed a method by computing the cross-correlation coefficient based similarity of an image with manually defined template images. Although high accuracy was reported, manual generation

of those template images from a large training image database was time-consuming and observer-dependent. Lehmann et al. [15] proposed a method using down-scaled image pixels with four distance measures along with K-nearest neighbor (KNN) classifier. Almost equal accuracy was reported when compared with the method of Arimura et al. [14] on their test set. Boone [2] developed a method using a neural network (NN) classifier working on down-sampled images. Recently, Luo [1] proposed a method containing two major steps including region of interest (ROI) extraction, and then classification by the combination of a Gaussian mixture model classifier and a NN classifier using features extracted from ROI. An accuracy of 98.2% was reported on a large test set of 3100 images. However, it was pointed out by the author that the performance of the method depended heavily on the accuracy of ROIs segmentation. Inaccurate or inconsistent ROI segmentations would introduce confusing factors to the classification stage. All the aforementioned work regarded the chest view identification task as a two class classification problem, however, we included an additional OTHER class in this work. The reason is that in order to build a fully automatic system to be integrated into CAD/PACS for identification of PA–AP and LAT chest radiographs, the system must filter out radiographs containing anatomy contents other than chest. Our task, therefore, becomes a three-class classification problem, i.e., identifying images of PA–AP, LAT, and OTHER, where “OTHER” are radiographs of head, pelvis, hand, spine, etc. In the more broad research field of object detection and recognition, many methods based on the use of local features have been proposed. The objects of interest were in many cases face, cars or people [16]–[21]. Cristinacce and Cootes [16] combined boosted detector [22] with the statistical shape model [23]. Multiple hypotheses of each local feature were screened using the shape model and the winning hypothesis was determined for each feature. Agarwal et al. [17] presented an object detection

340

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

Fig. 3. Images from the IRMA/ImageCLEF2008 database with the IRMA code annotated as: acquisition modality “overview image;” body orientation “AP unspecified;” body part “pelvis;” biological system “musculoskeletal.” Note the very high appearance variability caused by artifacts, diseases/implants, and different FoVs.

algorithm for detecting the side view of a car in a cluttered background. It used a “part-based representation” for the object. The global shape constraint was imposed through learning using the Sparse Network of Winnows architecture. Mohan et al. [20] proposed a full-body pedestrian detection scheme. They first used separate SVM classifiers to detect the body parts, such as heads, arms and legs. Then, a second SVM classifier integrating those detected parts was used to make the final detection decision. Leibe et al. [21] proposed a method for robust object detection based on learned codebook of local appearances. To integrate the global shape prior, an implicit shape model was learned to specify the locations, where the codebook entries might occur. Our work was inspired by many ideas from the nonmedical domain, but with more suitable models of human anatomy, accommodating the fact that in the medical domain “abnormality is the norm.” B. Proposed Approach We adopt a hybrid approach based on local feature detection and aggregation, followed by a global appearance check. It combines the use of discriminative, learning-based local-feature detectors, a filtering step based on localized shape prior constraints, and an exemplar-based global appearance check mechanism. The aim is to push the last few percentage points in performance gain, and to achieve high robustness against variations caused by diseases, imaging artifacts, occlusion, or missing data. These factors are common in clinical settings, and can severely alter the target anatomy to be difficult to recognize as a whole [see Fig. 1(b) and (c)] for computer algorithms. However, in most cases, at least a small portion of the anatomy (e.g., part of the lungs) will remain functional and recognizable. Therefore, focused, local approach is key to success. Because of the highly variable (and non-Gaussian) appearance of each local pattern (see Figs. 2 and 3), we choose statistical learning-based discriminative algorithm ([24]) over classical linear methods such as, for example, active shape and appearance model [25] which assumes Gaussian distributions . We also use local models to impose shape constraints; and we use a large, redundant number of them in order to achieve robustness. The key idea is to build many parametric models among small, overlapping groups of local features. These models, when statistically combined, can easily reveal and filter out errors in local feature detection (e.g., mistaking a large mass in the abdomen for heart). In the rare cases when the previous two steps fail to confidently make a decision, a third classification step is imposed, this time, based on a completely orthogonal and complemen-

tary philosophy: global as opposed to local, nearest neighbor in the pixel space instead of discriminative learning machines trained in a feature space. In our experiments, only 6% of cases passed through this last step which, nevertheless, did provide a solid last-stage performance gain. II. METHODS Fig. 4 shows an overview of the algorithm. Our algorithm is designed to first detect multiple focal anatomical structures within the medical image. This is achieved through a learning-by-example landmark detection algorithm that performs simultaneous feature selection and classification at multiple image scales. A second step is performed to eliminate inconsistent findings through a robust sparse spatial configuration (SSC) algorithm, by which consistent and reliable local detections are retained while outliers are removed. Finally, a reasoning module assessing the filtered findings, i.e., remaining landmarks, is used to determine the final content/orientation of the image. Depending on the classification task, a post-filtering component using the exemplar-based global appearance check for cases with low classification confidence may also be included to reduce false positive (FP) identifications. A. Landmark Detection The landmark detection module in this work was inspired by the work of Viola and Jones [22], but modified to detect an anatomic landmark point with variable contexts (e.g., the carina of trachea) instead of an image patch with a fixed region of interest (e.g., a face). The basic idea is to cast the detection problem as a feature selection and classification problem, and solve it using the AdaBoost algorithm proposed by [24]. Specifically, multiple classifiers, each of which is referred to as a “landmark detector,” were generated through supervised learning. Then each detector was used to determine whether a specific anatomic landmark was presented within an image. We developed a coarse-to-fine multiscale implementation to achieve real-time recognition, i.e., providing immediate response to support user-in-the-loop use scenarios. To train landmark detectors, we first collect a number of training images along with annotated anatomic positions (as shown in Fig. 5). Then, resampled images aligned at these positions are cropped and collected as positive training samples. A set of extended Haar features are computed as matching responses to various templates as shown in Fig. 6. The size of sub-patch samples for a specific detector is manually determined according to the landmark’s position in the image, and this size is fixed across all images at one scale. The sub-patch

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

341

Fig. 4. The overview of our approach for automatic medical image annotation: it starts with a landmark detection module to detect local anatomic structures. The detected landmark sets are then verified by a sparse spatial configuration algorithm. After that, the filtered landmark sets are converted to image classification scores. By comparing the maximum score to a high threshold and a low threshold, the image is determined as a specific class or OTHER. If the maximum score is between the two thresholds, the image will further go through a global appearance check module to make the final decision.

Fig. 5. Landmark annotation/detection example: For a PA–AP chest image, 12 anatomic positions are selected and annotated by radiologists. They are represented by crosses with underlying labels specifying landmark indices.

Fig. 6. Exemplar 2-D Haar templates. Haar features are computed by subtracting the sum of pixels in the grey rectangle(s) from the sum of pixels in the black rectangle(s). The outmost bounding square for a template corresponds to the sub-patch size for a detector.

size for different detectors varies from 13 13 pixels to 25 25 pixels. The cropped sub-patches are allowed to extend beyond the image border to less than 50% of the patch size, in which case the part of the patch falling outside the image is padded with zeroes. Generally, detectors for fine scale images have larger patch sizes than those for coarse scale images. The total number of Haar features computed for a detector varies

depending on its sub-patch size. For a detector with the size of 25 25 pixels, about 15 000 features are computed. Regarding the classifier, we employ the AdaBoost algorithm [22], [24] for simultaneous feature selection and classification. For each cascade level of a detector, the training criterion is to achieve high true positive rate (99%) and moderate false positive deduction rate on the training set. For the first level cascade, negative training sets are collected by randomly cropping sub-patches in images belonging to negative classes, and sub-patches far away from annotated positions.1 For subsequent levels, the negative training sets are obtained by collecting false positives generated using the partially trained detector. The entire training process stops when the ratio between the size of negative set generated using the current detector and the size of positive set is less than 0.1. A typical trained detector has 6–8 levels. Different landmark detectors are trained independently across several scales within down-scaled images. For this image annotation application, two scales with downscaling ratio of 1/8 (fine scale) and 1/16 (coarse scale) of the original image size are adopted to balance the computational time and detection accuracy. Important parameter settings for different modules in our approach are summarized in Table I. The whole detection procedure during testing phase is illustrated in Fig. 7. Firstly, the coarse scale landmark detector slides within the image, computes selected features and generates a response/detection score at each pixel position. Multiple candidate positions with responses larger than the detection threshold (determined automatically during training for the last level cascade) are then selected. After that, the landmark detector at finer scale is scrutinized at previously determined candidate position(s) to locate the local structure more accurately. A single position with the highest response (larger than the detection threshold for the last level cascade) is obtained as the final detection for a detector. The final outputs of a landmark detector 1The distance between the center of a negative sub-patch and the annotated position should be larger than 50% of the width/height for the sub-patch.

342

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

TABLE I SUMMARYOF PARAMETERS FOR DIFFERENT MODULES IN THE PROPOSED APPROACH

Fig. 7. Illustration of the landmark detection procedure: To detect the position of the left lung tip (LLT), the coarse scale detector first scans on the coarse-scale image. It detects several candidate positions (shown as crosses) with high responses. Next, the fine scale detector runs in the fine-scale image, within the focal sub-window(s) (shown as rectangles) centered in previously detected candidate position(s). A single position with the highest response is determined to be the final detection, i.e., LLT.

are the horizontal and vertical coordinates in the image along with a response/detection score. B. Sparse Spatial Configuration Algorithm Knowing that the possible locations of anatomical landmarks are rather limited, we aim to exploit this geometric property to

eliminate error detections from the first step. This geometric property is represented by a spatial constellation model among detected landmarks. The evaluation of consistency between a landmark and the model can be determined by the spatial relationship between the landmark and other landmarks, i.e., how consistent the landmark (as a candidate) is according to other landmarks. In this work, we propose a local filtering algorithm

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

(Alg. 1) to sequentially remove false detections until no outliers exist. Algorithm 1 Sparse Spatial Configuration Algorithm for each candidate

do

for each voting group of do

generated from the combinations

Compute the likelihood score

of

coming from

343

is the transformation matrix learned by linear regreswhere is the matrix formed by the sion from a training set, and x–y coordinates of landmarks in the voting group. A high likemeans that the candidate is likely lihood score of to be a good local feature detection according to its spatial rela. tions with other landmarks in Here we briefly illustrate the learning procedure for the transin (2). Without loss of generality, we deformation matrix scribe the procedure for three landmarks annotated in training images

end for Sort all the scores received by ). denoted as

. (The sorted array is

end for

where and stand for the horizontal and vertical positions image for the landmark candidate . Assuming that of the landmark 2 and landmark 3 form one voting group for landmark 1, we try to predict the horizontal position of landmark 1 (denoted as ) using the voting group. The spatial relationship beand positions of landmark 2 and landmark 3 (denoted tween as and respectively) is modeled using linear regression as

repeat

if

then Remove

and all scores involved with .

end if until No more candidate are removed In general, our reasoning strategy sequentially “peels away” erroneous detections from the landmark set (denoted as X) of one anatomy class. We denote the th detected landmark as , which is a 2-D variable with values corresponding to the detected x–y coordinates in the image. Each candidate receives a set of likelihood scores generated from its spatial relationship with voting groups formed by other landmarks. Multiple voting groups are generated by the combination of different (denoted as landmarks from the landmark set X excluding ). The size of each voting group is designed to be small, and the combinations with size of 1–3 are used in this work. In this case, if there are 10 landmarks of one anatomy class detected in voting groups would be the image, a total of formed for each candidate. The advantage of this setting lies in that even when there are many missing or erroneous detections, there exists a sufficient number of “good” voting groups to be formed from “good” detections. These groups are then used to effectively remove bad detections. In this sense, the sparse and distributed nature (within voting groups) of the algorithm guarantees that the shape prior constraint could still take effect even in handling challenging cases, such as those with a large percentage of occlusion, or missing data. from its The likelihood score received by candidate voting group is modeled as multivariant Gaussian as following: (1) is the estimated covariance matrix, and where predicted position coming from modeled as

is the (2)

(3) where is additive Gaussian noise. The vertical position of landmark 1 (denoted as ) could be modeled similarly. Formulating variables in a vector format as

The transformation matrix can be derived using maximum likelihood estimation method [26] as (4) The reasoning strategy (Alg. 1) then iteratively determines whether to remove the current “worst” candidate (denoted as ), which is the one receiving the smallest maximum score compared with other candidates. The algorithm will remove the “worst” candidate if its score is smaller than a predefined threshold . This process will continue until no outlier exists. See Section III-C2 for quantitative experimental results and examples of false positive detections filtered out by this step. The proposed SSC algorithm is different from previously proposed part-based recognition methods (e.g., [20], [27]) in the following respects: Firstly, the algorithm explicitly models the shape priors and spatial relationships between part-based detections (i.e., anatomic landmarks). The shape priors are constructed via an automatic learning procedure, instead of being heuristically confined (e.g., [20]). Secondly, the proposed SSC

344

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

algorithm has an iterative outlier filtering mechanism using multiple voting groups formed by small numbers of landmarks. This makes the method more flexible and effective in handling a potentially large percentage of missing and erroneous detections, and thus more robust in recognizing challenging images.

to have the longest edge of 512 pixels while preserving the aspect ratio. 1) The in-house database includes images collected from daily imaging routine from radiology departments in hospitals. It contains a total of 9789 radiographs including 5875 chest radiographs and 3914 other radiographs from a variety of other anatomy classes. The image class labels were provided and verified by radiologists. In this database, the chest images cover a large variety of chest exams and diseases, representing image characteristics from real world PACS. It includes upright position from normal patients, supine position of critically ill patient in intensive or emergency care units, and a small number of pediatric images with both suspended and lying position. In addition, the image quality ranges from decent contrast and well-set gray level to low contrast images, or images with severe pathology or implants. In this work, the OTHER class excluded radiographs of certain anatomies (e.g., radiographs of finger and shank) with large difference of the image aspect ratio in comparison to chest radiographs. These radiographs could be easily differentiated from chest radiographs using heuristic rules based on the aspect ratio. We randomly selected 500 PA–AP, 500 LAT, and 500 OTHER images for training landmark detectors. These images were also used as the exemplar image database for the post-filtering component. The remaining 8289 images were used as the testing set. 2) For the multiclass medical radiograph annotation task, we used the IRMA/ImageCLEF2008 database [28]. All images from this database have been labeled with a detailed code that specified acquisition modality, body orientation, body part and biological system. It contains more than 10 000 images from total 197 unique classes. The distribution of different classes in this database is not uniform. For example, the chest radiographs comprise about 37% of the total images. And the top nine classes comprise about 54% of the total images. In this work, we selected a subset of images (the top nine classes with the most number of images) from this database, including PA–AP chest, LAT chest, PA–AP left hand, PA–AP cranium, PA–AP lumbar spine, PA–AP pelvis, LAT lumbar spine, PA–AP cervical spine, and LAT left to right cranium. The remaining images were regarded as one OTHER class. For the PA–AP and LAT chest images, we directly used the detectors trained using the in-house database. 50 PA–AP and 50 LAT chest testing images were randomly selected from the IRMA/ImageCLEF2008 database. For the remaining seven classes, we randomly selected 200 images for each class. 150 images were used as the training set, and the remaining 50 images were used for testing. For the OTHER class, we randomly selected 2000 training and 2000 testing images each. Table II summarizes the images used in different classification tasks. We added an additional number of rotation modified training images to improve the landmark detectors’ robustness against image rotation. Specifically, 150 randomly selected images for each class (except the OTHER class) from the training , and were set were randomly rotated within the range of

C. Classification Logic For an image belonging to a certain anatomy class, we assume that there would be a sufficient number of true positive landmarks to be detected. Therefore, the classification logic is determined as follows: the number of verified landmarks for each image class is divided by the total number of detectors for that class, representing the final classification score (denoted as for the th image class). The image class with the maximum classification score is then selected as the candidate class. In case that equal classification scores are obtained between several classes, the class with the maximum average landmark detection score is chosen as the candidate class. The classification decision is then made by comparing the classification score of the candidate class with predefined thresholds (see Fig. 4 for illustration). Depending on the classification task, a FP reduction module based on the global appearance check may also be used for those images with low classification confidence, i.e., images . with classification scores within the defined range of A large portion of these images come from the OTHER class. They have a small number of local detections belonging to the candidate image class, yet their spatial configuration is strong enough to pass the SSC stage. Since previous local detection integration steps could not provide sufficient discriminative information for classification, we try to integrate a post-filtering component based on the global appearance check to make the final decision. In our experiment for PA–AP/LAT/OTHER separation task, only about 6% of cases went through this stage. To meet the requirement for real-time recognition, an efficient exemplar-based global appearance check method is adopted. Specifically, we use pixel intensities from 16 16 down-scaled image as the feature vector along with 1NN, which uses the Euclidean distance as the similarity measurement. With the complementary global appearance information, the FP reduction module could effectively remove FP images coming from the OTHER class, thus leading to the overall performance improvement of the final system (see Section III-B). III. EXPERIMENTS AND RESULTS A. Datasets In this work, we ran our method on four subtasks: PA–AP/LAT chest image view identification task with and without OTHER class, and the multiclass medical image annotation task with and without OTHER class. For the chest image identification task, we used a large-scale in-house database, and for the multiclass radiograph annotation task, we used the IRMA/ImageCLEF2008 database2. For radiographs with 12-bits or more intensity levels, we reduced their intensity levels to 8 bits. And all images in these databases were down-scaled 2http://imageclef.org/2008/medaat

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

345

TABLE II IMAGES USED FOR DIFFERENT CLASSIFICATION TASKS

TABLE III PA–AP/LAT/OTHER CHEST RADIOGRAPHS ANNOTATION PERFORMANCE

combined with the original training set. For the multiclass annotation task, there were a total of 300 (150 original and 150 rotated) images for each of the remaining seven class (except chest) and 2000 images for the OTHER class to train landmark detectors. B. Classification Performance For the chest radiograph annotation task, we compared our method with three other methods described by Boone et al. [2], Lehmann et al. [15], and Kao et al. [13]. For the method proposed by Boone et al. [2], we down-sampled the image to the resolution of 16 16 pixels and constructed a five hidden nodes NN. For the method proposed by Lehmann et al. [15], a five nearest neighbor (5-NN) classifier using 32 32 down-sampled image with the correlation coefficient distance measurement was used. The same training image database (500 PA–AP/500 LAT/500 OTHER) for the postfiltering component of our method was used for the 5-NN classifier. For the method proposed by Kao et al. [13], we found that the projection profile derived features described in the literature were sensitive to the orientation of anatomy and noise in the image. Directly using the smoothed projection profile as the feature along with the LDA classifier provided better performance. Therefore, we used this improved method as our comparison.

TABLE IV MULTICLASS RADIOGRAPHS ANNOTATION PERFORMANCE

For the multiclass radiograph annotation task, we compared our method with the in-house implemented BOF method proposed by Deselaers and Ney [4] (named as PatchBOF+SVM) and the method proposed by Tommasi et al. [9] (named as SIFTBOF+SVM). Regarding PatchBOF+SVM, we used the BOF approach based on randomly cropped image sub-patches. The generated BOF histogram for each image had 2000 bins, which were then classified using a SVM classifier with a linear kernel. Regarding SIFTBOF+SVM, we implemented the same modified version of SIFT (modSIFT) descriptor and used the same parameters for extracting BOF as those used by Tommasi et al. [9]. We combined the 32 32 pixel intensity features and the modSIFT BOF as the final feature vector, and we used a SVM classifier with a linear kernel for classification. We also tested the benchmark performance of directly using 32 32 pixel intensity from the down-sampled image as the feature vector along with a SVM classifier. Tables III and IV summarize the recognition rates of various methods using the whole training set and the testing set. It can be seen that our approach obtained a very high accuracy on the PA–AP/LAT separation task, and it also performed the best for the other three tasks. To further analyze various algorithms’ performance and stability on different datasets, we ran multiple rounds of training combined with bootstrapping test. Specifically, the training set was randomly sampled to select 70% samples for ten times, and thus a total of 10 training sets were created; associated with each sampled training set, 1000 testing sets

346

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

TABLE V PA–AP/LAT/OTHER CHEST RADIOGRAPHS ANNOTATION PERFORMANCE FOR MULTIPLE ROUNDS OF TRAINING AND CORRESPONDING BOOTSTRAPPING TEST

we consulted the book by Netter [29]. To test the landmark detectors’ performance, we annotated 100 PA–AP and 100 LAT chest images separately. Since the landmark detectors run on the down-scaled images, the detected position could deviate from the ground truth position to certain degree, which is allowable for our image annotation application. We determine the detected landmark as true positive detection when the distance between the detected position and the annotated ground truth position is smaller than 30 pixels. Note that the detection performance can be traded off against computational time. Currently in order to achieve real-time performance, we accepted an average sensi, tivity for the 23 chest landmark detectors at 86.91% which was good enough to support the aforementioned overall system performance. 2) SSC: For a PA–AP/LAT separation task on 200 images with annotated ground truth of landmarks, 55 out of 356 false positive landmark detections were filtered by the SSC algorithm, while the true positive detections were unaffected. In addition, the algorithm removed on average 941 and 486 false positive detections for the PA–AP/LAT/OTHER task and the multiclass task with OTHER class. Fig. 9 shows that the result of the SSC algorithm in reducing false positive detections on nonchest image classes.

TABLE VI MULTICLASS RADIOGRAPHS ANNOTATION PERFORMANCE FOR MULTIPLE ROUNDS OF TRAINING AND CORRESPONDING BOOTSTRAPPING TEST

D. Failure Case Analysis

were randomly generated by bootstrapping (sampling with replacement) using the original testing set. Therefore, a total of 10 000 testing results were obtained for each method. Average precisions with corresponding confidence intervals at 0.95 of various methods were computed and summarized in Tables V and VI. The performance differences between our algorithm and other algorithms were found to be statistically significant for all the tasks, according to the two-sided Wilcoxon (rank sum) test with confidence interval at 0.95. Fig. 8 shows the classification result along with the detected landmarks for different classes. It is seen that the learned landmark detectors were robust to scale variance and medium degree of image rotation (less than ) as shown in the examples of hand images in the second row of Fig. 8. The proposed annotation method can robustly recognize challenging cases under the influence of strong artifacts or severe diseases. C. Intermediate Results In this section, we provide intermediate results and performance numbers for both the landmark detection and the SSC module. 1) Landmark Detection: The landmark detection procedure is invariant to image translation since detectors are scanned on the whole image. The rotation and scale robustness is implicitly achieved through the training procedure. In this work, we used 12 landmarks for PA–AP and 11 for LAT chest images. For additional image classes in the multiclass annotation task, we used 7–9 landmarks for each class. For the selection of landmarks

For the PA–AP/LAT annotation task, the majority of failed cases were pediatric PA–AP images as shown in Fig. 10(a) and (b). This may be caused by that the training chest radiographs were comprised predominantly of images from nonpediatric patients. As a result, the image patterns in these examples were inconsistent for both the trained detectors and the 1NN classifier, and thus caused the classification error. Regarding the FP images from the OTHER class for the multiclass annotation task, the majority of failures were due to insufficient coverage of anatomy objects or severe occlusion of artifacts [as shown in Fig. 10(c) and (d)]. These cases had a few FP landmarks detected, and they went through the global appearance check stage. However, because the appearances of these images were not similar to a typical image from the OTHER class, they were misclassified by the 1NN classifier. IV. DISCUSSION A. Computation Complexity The training time for a landmark detector varied depending on the patch size and number of training images. For the PA–AP/ LAT/OTHER classification task, it took about 4 hours to train a detector with the patch size of 25 25 pixels on Intel Xeon 1.86 GHz with 3.00 GB RAM. Regarding the testing phase, the computation complexity of our method comes mainly from the landmark detection procedure. Although the cascade classification framework guarantees the run time efficiency for each landmark detector, the computation cost increases linearly with the number of specified landmarks and the number of image classes. To meet the requirement for online recognition, one possible solution is to use a few coarse resolution landmark detectors to first select several candidate image classes, and then the multiscale

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

347

Fig. 8. Examples of the detected landmarks on different images. The anatomy classes of these images were correctly recognized by the proposed algorithm in the experiments.

detectors from the selected candidate classes are used to determine the final class. In this work, this scheme was adopted for the multiclass medical radiograph annotation task. According to our experiments, based on a multithread implementation of the algorithm, the entire process time on average for an image was about 1 s for the PA–AP/LAT/OTHER classification task and 2 s for the multiclass task. This satisfies our requirement for the online recognition procedure. B. System Extension The proposed algorithm framework is generalizable, and it could be applied to other image modalities and extended for other image parsing applications beyond radiograph annotation. In this work, we further exploit the by-products of the algorithm, more specifically, the detected/filtered landmarks on chest images, for two other image parsing applications.

They are anatomy/organ ROI prediction and optimized image visualization. 1) Anatomy/Organ ROI Prediction: From the filtered remaining landmark set , we could define the organ-specific anatomic ROI, which is represented by a bounding rectangle (vertices) of the bounding in the image. The parameters rectangle from the ROI are computed by (5) where is a ROI prediction function learned by linear regression similarly as in (3). Fig. 11 shows examples of the predicted ROI of lung, abdomen, and heart in topograms. As can be seen in Fig. 11(c) and (d), even with incomplete local detections, the algorithm can still robustly predict ROIs for abnormal images, e.g., images with the FoV variation and disease. With the recognition of anatomical contents, other organ-specific algorithms

348

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

Fig. 9. The SSC algorithm performance on different image classes (better viewed in color): (a) LAT chest, (b) foot, (c) cranium, and (d) hand. The blue colored crosses are true positive landmark detections; the yellow colored ones are false positive detections; and the red colored ones are detections filtered by the SSC algorithm. APPA and LAT label under the detected landmarks specify that detections are from PA–AP chest detectors or LAT chest detectors.

Fig. 10. Examples of failure images: (a) and (b) pediatric PA–AP chest images were misclassified as “LAT chest,” (c) and (d) images from the OTHER class with insufficient coverage of anatomy objects were misclassified as “Pelvis” and “LAT chest,” respectively.

Fig. 11. Anatomy ROI prediction in the topogram for CT scan automation. The blue rectangle bounded regions in (a)–(d) correspond to ROIs of lung, abdomen, heart, and lung. It can be seen that the ROI prediction algorithm is robust under the FoV variation and disease influences.

(e.g., a lung nodule detection algorithm) can be invoked on a subregion in the image to improve the specificity and efficiency for the final CAD system. 2) Optimized Image Visualization: The patient may stoop sometimes when taking the LAT chest image, and this may cause the image to present certain degrees of tilt as shown in Fig. 2(c) and (d) compared with Fig. 2(a), which is an image with a standard upright body position. We could explore the detected landmarks on the LAT chest image to perform registration with images with standard upright position. This allows robust online orientation correction for chest radiographs for optimized image visualization. Compared with previously proposed methods [11], [30], [31], where the orientation correction is restricted to images with rotations of only 90 , 180 , 270 , our system is more flexible and robust in estimating degrees of tilt. The orientation corrected images have the potential to help radiologists to view the LAT chest radiograph more conveniently. In addition, with the detected landmarks on PA–AP and LAT chest images, a synchronized view mechanism could be modeled to help radiologists to find the corresponding positions between the two images. More specifically, we first compute the relative position of the pinpointed cross on the PA–AP image [as shown

in Fig. 12(a)] within the frontal lung ROI defined in the PA–AP image. Assuming the relative position is roughly unchanged in the lateral lung ROI in the LAT image, we could estimate the corresponding position/range on the LAT image. Fig. 12 shows an example of the synchronized view feature of our visualization system. The band area (between the two solid lines) on the LAT image corresponds to the position pinpointed by the cross in the PA–AP image. This optimized display feature has the potential to help radiologists to locate and scrutinize the corresponding findings on PA–AP and LAT chest images simultaneously. V. CONCLUSION We have developed a hybrid learning-based approach for parsing and annotation of medical radiographs. Our approach integrates learning-based local appearance detections, the shape prior constraint by a sparse spatial configuration algorithm, and a final filtering stage with the exemplar-based global appearance check. This approach is highly accurate, robust, and fast in identifying images even when severely altered by diseases, implants, or imaging artifacts. The advantages of the proposed approach lie in several aspects.

TAO et al.: ROBUST LEARNING-BASED PARSING AND ANNOTATION OF MEDICAL RADIOGRAPHS

349

Fig. 12. Optimized image visualization: (a) The PA–AP chest image with the pinpointed position shown as cross, (b) and the orientation corrected LAT chest image with the estimated corresponding position/range shown within the band area (between the two solid lines). Figure (c) and (d) show that when the pinpointed position on the PA–AP chest image moves, the corresponding position on the LAT chest image moves accordingly.

• The algorithm accurately detects semantic local visual cue representations (i.e., landmarks) in real-time. These landmark detectors generate a concise codebook representation which, acting together, also normalizes transformations and geometrical variations in translation, scale, and rotation. Compared with the popular BOF approaches (e.g., [4], [9]), spatial anatomical location is preserved in our model, and this is beneficial in at least two ways: first, each classification task is easier; second, shape priors can be learned and enforced. • Our shape prior modeling module is based on a group of learned sparse spatial configuration models. This step enforces the spatial anatomical consistency of local findings. Because of its sparse nature, i.e., it is a collection of spatial relations among many small groups of landmarks, the shape prior constraint could still take effect even with many missed detections. Compared with methods using global shape representations (e.g., [25], [32]), our algorithm can be particularly effective on challenging cases with a large percentage of occlusion, or missing data, such as cases with large tumor or liquids in the lungs. • The additional global appearance check step further improved the final performance (from 98.08% to 98.49%), by providing complementary information that is not fully captured by previous integrated local detections. It may seem that the percentage gains in discussion here are not large, however the improvement in users’ experiences in the clinical environment is quite dramatic: for a busy clinic, the difference above represents one error several weeks versus one error several days. The experimental results on a large-scale chest radiograph view position identification task and a multiclass medical radiograph annotation task have demonstrated the effectiveness and efficiency of our method. As a result, minimum manual intervention is required, improving the usability of such systems in the clinical environment. Due to the generality and scalability of our approach, it has the potential to be extended in several directions, for example, annotation of more classes of radiograph images, extensions to other imaging modalities, and 2-D/3-D ROI detection tasks. Fig. 11 shows the example application of the automatic organ ROI prediction in topogram for CT scan automation [33], [34]. In addition, the same algorithm framework has been extended for 3-D medical image applications, e.g., coronary artery detection and tracing in CT image [35].

In summary, this paper presents a robust and generalizable algorithm framework for medical radiograph annotation. Extensive experiments were conducted to validate the system performance both in terms of accuracy and speed. Such systems can dramatically improve radiology workflow and save valuable time and cost in the clinical environment. ACKNOWLEDGMENT The authors would like to express their gratitude to M. Freedman, B. Jian, and Y. Zhan for their valuable discussions and comments on this paper. REFERENCES [1] H. Luo, W. Hao, D. Foos, and C. Cornelius, “Automatic image hanging protocol for chest radiographs in PACs,” IEEE Trans. Inf. Technol. Biomed., vol. 10, no. 2, pp. 302–311, 2006. [2] J. M. Boone, G. S. Hurlock, J. A. Seibert, and R. L. Kennedy, “Automated recognition of lateral from PA chest radiographs: Saving seconds in a PACs environment,” J. Dig. Imag., vol. 16, no. 4, pp. 345–349, 2003. [3] T. Deselaers, T. M. Deserno, and H. Müller, “Automatic medical image annotation in imageclef 2007: Overview, results, and discussion,” Pattern Recognit. Lett., vol. 29, no. 15, pp. 1988–1995, 2008. [4] T. Deselaers and H. Ney, “Deformations, patches, and discriminative models for automatic annotation of medical radiographs,” Pattern Recognit. Lett., vol. 29, no. 15, pp. 2003–2010, 2008. [5] U. Avni, H. Greenspan, M. Sharon, E. Konen, and J. Goldberger, “X-ray image categorization and retrieval using patch-based visual words representation,” in Proc. IEEE Symp. Bio-Med. Imag. (ISBI), 2009, pp. 350–353. [6] H. Müller, T. Gass, and A. Geissbuhler, “Performing image classification with a frequency-based information retrieval schema for imageclef 2006,” Proc., ImageCLEF 2006, Ser. Working Notes of the Cross Language Evalutation Forum (CLEF 2006), 2006. [7] M. O. Güld and T. M. Deserno, “Baseline results for the imageCLEF 2007 medical automatic annotation task using global image features,” Adv. Multilingual Multimodal Inf. Retrieval, vol. 4730, pp. 637–640, 2008. [8] D. Keysers, T. Deselaers, C. Gollan, and H. Ney, “Deformation models for image recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 29, no. 8, pp. 1422–1435, Aug. 2007. [9] T. Tommasi, F. Orabona, and B. Caputo, “Discriminative cue integration for medical image annotation,” Pattern Recognit. Lett., vol. 29, no. 15, pp. 1996–2002, 2008. [10] D. Lowe, “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vis., vol. 60, no. 2, pp. 91–110, 2004. [11] M. Evanoff and K. McNeill, “Automatically determining the orientation of chest images,” Proc. SPIE, vol. 3035, pp. 299–308, 1997. [12] E. Pietka and H. K. Huang, “Orientation correction for chest images,” J. Dig. Imag., vol. 5, no. 3, pp. 185–189, 1992. [13] E. Kao, C. Lee, T. Jaw, J. Hsu, and G. Liu, “Projection profile analysis for identifying different views of chest radiographs,” Acad. Radiol., vol. 13, no. 4, pp. 518–525, 2006.

350

IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 30, NO. 2, FEBRUARY 2011

[14] H. Arimura, S. Katsuragawa, T. Ishida, N. Oda, H. Nakata, and K. Doi, “Performance evaluation of an advanced method for automated identification of view positions of chest radiographs by use of a large database,” Proc. SPIE Med. Imag., vol. 4684, pp. 308–315, 2002. [15] T. M. Lehmann, O. Güld, D. Keysers, H. Schubert, M. Kohnen, and B. B. Wein, “Determining the view of chest radiographs,” J. Dig. Imag., vol. 16, no. 3, pp. 280–291, 2003. [16] D. Cristinacce and T. Cootes, “Facial feature detection using adaboost with shape constraints,” in Proc. Br. Mach. Vis. Conf. (BMVC), 2003, pp. 231–240. [17] S. Agarwal, A. Awan, and D. Roth, “Learning to detect objects in images via a sparse, part-based representation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 26, no. 11, pp. 1475–1490, Nov. 2004. [18] K. Yow and R. Cipolla, “Feature-based human face detection,” Image Vision Comput., vol. 15, no. 9, pp. 713–735, 1997. [19] T. K. Leung, M. C. Burl, and P. Perona, “Finding faces in cluttered scenes using random labeled graph matching,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 1995, pp. 637–644. [20] A. Mohan, C. Papageorgiou, and T. Poggio, “Example-based object detection in images by components,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 4, pp. 349–361, Apr. 2001. [21] B. Leibe, A. Leonardis, and B. Schiele, “Robust object detection with interleaved categorization and segmentation,” Int. J. Comput. Vis., vol. 77, no. 1, pp. 259–289, 2008. [22] P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” in Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), 2001, vol. 1, pp. 511–518. [23] I. Dryden and K. V. Mardia, The Statistical Analysis of Shape. New York: Wiley, 1998. [24] Y. Freund and R. Schapire, “A decision-theoretic generalization of on-line learning and an application to boosting,” J. Comput. Syst. Sci., vol. 55, no. 1, pp. 119–139, 1997.

[25] T. Cootes et al., “Active appearance models,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 6, pp. 681–685, Jun. 2001. [26] P. J. Huber, Robust Statistics. New York: Wiley, 1981. [27] M. Toews and T. Arbel, “A statistical parts-based model of anatomical variability,” IEEE Trans. Med. Imag., vol. 26, pp. 497–508, 2007. [28] T. Deselaers and T. Deserno, “Medical image annotation in imageCLEF 2008,” in Proc. CLEF Workshop 2008: Evaluating Syst. Multilingual Multimodal Inf. Access, 2009. [29] F. H. Netter, Atlas of Human Anatomy, Ser. Netter Basic Science, 4th ed. : Elsevier Health, 2006. [30] H. Luo and J. Luo, “Robust online orientation correction for radiographs in PACs environments,” IEEE Trans. Med. Imag., vol. 25, no. 10, pp. 1370–1379, Oct., 2006. [31] K. Nahm, “Automatic detection of the lung orientation in digital PA chest radiographs,” J. Opt. Soc. Korea, vol. 1, pp. 60–64, 1997. [32] M. Leventon, W. Grimson, and O. Faugeras, “Statistical shape influence in geodesic active contours,” in Proc. IEEE Int. Conf. Comput. Vis. Pattern Recognit. (CVPR), 2000, vol. 1, pp. 316–323. [33] Z. Peng, Y. Zhan, X. S. Zhou, and A. Krishnan, “Robust anatomy detection from CT topograms,” Proc. SPIE Med. Imag., vol. 7620, pp. 1–8, 2009. [34] Y. Zhan, X. S. Zhou, Z. Peng, and A. Krishnan, “Active scheduling of organ detection and segmentation in whole-body medical images,” Proc. of MICCAI, vol. 5242, LNCS, pp. 313–321, 2008. [35] L. Lu, J. Bi, S. Yu, Z. Peng, A. Krishnan, and X. S. Zhou, “A hierarchical learning approach for 3-D tubular structure parsing in medical imaging,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), Sep. 2009, pp. 2021–2028.

Robust Learning-Based Parsing and Annotation of ...

Feb 2, 2011 - *X. S. Zhou is with the Siemens Medical Solutions USA, Inc., Malvern, PA. 19355 USA (e-mail: ...... In ad- dition, the algorithm removed on average 941 and 486 false pos- .... The authors would like to express their gratitude to.

2MB Sizes 1 Downloads 202 Views

Recommend Documents

Robust Learning-Based Parsing and Annotation of ...
Feb 2, 2011 - Our algorithm was used to enhance advanced image visualization workflows by ... THE amount of medical image data produced nowadays.

Universal Dependency Annotation for Multilingual Parsing
of the Workshop on Treebanks and Linguistic Theo- ries. Sabine Buchholz and Erwin Marsi. 2006. CoNLL-X shared task on multilingual dependency parsing. In.

Robust Learning-Based Annotation of Medical ...
For this application, two scales are sufficient to .... OTHER class, thus leading to the overall performance improvement of the final system (see Section 3).

PartBook for Image Parsing
effective in handling inter-class selectivity in object detec- tion tasks [8, 11, 22]. ... intra-class variations and other distracted regions from clut- ...... learning in computer vision, ECCV, 2004. ... super-vector coding of local image descripto

algebraic construction of parsing schemata
Abstract. We propose an algebraic method for the design of tabular parsing algorithms which uses parsing schemata [7]. The parsing strategy is expressed in a tree algebra. A parsing schema is derived from the tree algebra by means of algebraic operat

On the Complexity and Performance of Parsing with ... - GitHub
seconds to parse only 31 lines of Python. ... Once these are fixed, PWD's performance improves to match that of other ...... usr/ftp/scan/CMU-CS-68-earley.pdf.

SpatialML: Annotation Scheme, Corpora, and Tools
statistical entity tagger and a disambiguator. Both these tools are built on top of the freely available Carafe11 machine learning toolkit. The entity tagger uses a. Conditional Random Field learner to mark up PLACE tags in the document, distinguishi

Capture and Parsing of HTTP Packets
Given a personal computer equipped with a packet capture library, write a program in C language that: • Captures all the packets generated and received by the host. • Writes, per each packet, a single line on screen reporting the following inform

Parsing words - GitHub
which access sequence elements without bounds checking (Unsafe sequence operations). ...... This feature changes the semantics of literal object identity.

Pfff: Parsing PHP - GitHub
Feb 23, 2010 - II pfff Internals. 73 ... 146. Conclusion. 159. A Remaining Testing Sample Code. 160. 2 ..... OCaml (see http://caml.inria.fr/download.en.html).

algebraic construction of parsing schemata
matics of Language (MOL 6), pages 143–158, Orlando, Florida, USA, July 1999. ... In Masaru Tomita, editor, Current Issues in Parsing Technology, pages ...

PartBook for Image Parsing
effective in handling inter-class selectivity in object detec- tion tasks [8, 11, 22]. ... automatically aligning real-world images of a generic cate- gory is still an open ...

Web Services Annotation and Reasoning
handle and to manipulate them by introducing Semantic Web technologies and additional logical formalisms into the annotation process. The annotation proc-.

Annotation Cribsheet Single.pdf
LEARNING. Use these headings to explain each piece of work you have done in your book. Page 1 of 1. Annotation Cribsheet Single.pdf. Annotation Cribsheet ...

XML and MPEG-7 for Interactive Annotation and ...
Security enhanced Distributed Knowledge Management architecture. Petros Belsis, ... on the expansion of knowledge sharing capabilities, others emphasizing on .... XML files. This choice enables operational interoperability, because the ... domains (i

zotero pdf annotation
Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. zotero pdf annotation. zotero

Posterior Sparsity in Unsupervised Dependency Parsing - Journal of ...
Department of Computer and Information Science ... We investigate unsupervised learning methods for dependency parsing models that .... this interpretation best elucidates how the posterior regularization method we propose in Section 4.

Importance of linguistic constraints in statistical dependency parsing
Importance of linguistic constraints in statistical dependency parsing. Bharat Ram Ambati,. Language Technologies Research Centre, IIIT-Hyderabad, India. Motivation. • Machine Translation. Indian Language Indian Language Indian Language English. In

parsing techniques pdf
Page 1 of 1. File: Parsing techniques pdf. Download now. Click here if your download doesn't start automatically. Page 1 of 1. parsing techniques pdf. parsing ...

Semantic Image Retrieval and Auto-Annotation by ...
Conventional information retrieval ...... Once we have constructed a thesaurus specific to the dataset, we ... the generated thesaurus can be seen at Table 5. 4.3.