OBJECT-BASED ULTRASOUND VIDEO PROCESSING FOR WIRELESS TRANSMISSION IN CARDIOLOGY Paul Rodriguez V*, Marios S. Pattichis, Constantinos S. Pattichis, Rony Abdallah, and Mary Beth Goens

1.

INTRODUCTION

The transmission of medical video over wireless channels requires the use of scalable Codecs, with error-resilient and error-concealment functionalities. A fundamental challenge in developing new methods for effective scalable coding of medical video is the careful and meaningful assessment of the coding performance. In this Chapter, we introduce a new, diagnostic approach to help guide the video compression process. A new segmentation method, based on multidimensional Amplitude-Modulation FrequencyModulation (AM-FM) is introduced for segmenting motion-mode ultrasound video into diagnostically meaningful objects. Rate-distortion curves are provided to provide a measure of video quality-scalability within the framework of object-based coding. There has recently been a number of studies assessing the effect of video compression on diagnostic accuracy (Ehler et al., 2000; Garcia et al. 2001; Karson et al., 1996; Main et al., 2000; Soble et al., 1998; Thomas et al., 2002). These clinical studies advocated the use of digital video technologies over video tapes. However, they did not present any new video compression methods, or even show the use of existing video compression technology that is appropriate for wireless transmission. In the image processing literature, effective video compression techniques are described as perceptually-lossless. This implies that no visual information is lost in the compression process. In medical image quality assessment, an effective test is to perform a blind test, asking the readers to select the better image between two unknown images (the compressed and uncompressed). If the readers select either video image an equal

*

Paul Rodriguez V. Department of ECE, Room 229-A, ECE Building, University of New Mexico, Albuquerque, NM 87131-1356, USA. [email protected] 491

492

P. RODRIGUEZ ET AL.

Figure 1. An Motion-mode (M-mode) video image. The thin white lines define boundaries between objects of interest. In the upper rectangular region, patient information has been deleted.

number of times, or if the compressed images are selected more often than the uncompressed images, then the compression scheme is deemed effective. The compression assessment problem is even more difficult in wireless video compression. The requirement to achieve very low bitrates leads to visual artifacts that are often unacceptable. Also, if high compression ratios are required, the requirement to achieve perceptually-lossless compression maybe unattainable. Clearly though, we may be able to accept visual artifacts if they have no measurable impact on clinical practice. We are thus led to introduce a new method for assessing video quality that we will call diagnostically-lossless. We assume that human readers are used for obtaining diagnosis. Then, perceptually-lossless videos are also diagnosticallylossless. Thus, we expect that diagnostically-lossless video compression can lead to improved video compression rates since the diagnostically-lossless video compression ratio will be at least as high as the one achieved in perceptually-lossless compression. The task of defining diagnostically-lossless video compression is very challenging. To recognize the challenges, consider the case where a sonographer obtains videos at a remote site, and then transmits them wirelessly for diagnostic assessment. The basic idea is to adopt a protocol that will allow effective diagnostic assessment after the video has been compressed. Yet, at the time of compression, there is no way of assessing all possible diagnostic outcomes.

OBJECT-BASED ULTRASOUND VIDEO

493

I

II

III

IV

V Figure 2. M-mode video image partition into text objects and the main video image.

As we shall describe in detail, the development of diagnostic compression methods requires the use of effective segmentation methods. We introduce a new, multidimensional AM-FM approach that can simultaneously track all the cardiac walls of diagnostic interest. In Section 2, we provide a summary of the use of digital video in echocardiography. This is followed by a summary of object-based video compression in Section 3. In Section 4, we introduce a new model for diagnostically-lossless compression of M-mode. A novel AM-FM representation for M-mode ultrasound is discussed in Section 5. We use the AM-FM model to derive a segmentation system for video objects of interest. The results are given in Section 6, and concluding remarks are summarized in Section 7.

2. DIGITAL VIDEO TRANSMISSION IN CARDIOLOGY A summary of video compression technology is beyond the scope of this Section, and can be found in an earlier Chapter of this book by Pattichis et al. (2005). In this Section, we summarize clinical studies that studied the use of digital video in echocardiography. None of these studies looked into the possibility of developing a video compression software system based on the concept of diagnostically lossless compression. Instead, they provided a summary of testing existing technologies. The use of digital video in echocardiography has been studied at-least since 1996, as discussed by Karson et al (1996). In 1996, Karson et al established that a 20:1 JPEG compression for still images did not degrade diagnostic accuracy. Soble et al. and Garcia et al. (Soble et al., 1998; Garcia et al., 2001) argued that MPEG-1 video delivered the same diagnostic accuracy as the super VHS format. Real-time transmission of MPEG-2

494

P. RODRIGUEZ ET AL.

over ATM at 2 Mbps was studied by Main et al. (2000). A multi-site study was reported by Ehler et al. (2000). More recently, Thomas et al. (2002) argued for the wide adoption of digital echocardiography based on the DICOM standard.

3. A SUMMARY OF OBJECT-BASED VIDEO COMPRESSION Object-based coding has been recently adopted by the MPEG-4 video compression standard (ISO/IEC., 1999, IS 14496-1; ISO/IEC., 1999, IS 14496-2, ISO/IEC., 1999, IS 14496-X). It is widely accepted that effective video compression at very low bitrates can only be achieved through object-based coding. This approach allows us independent rate control on different parts of the video. Yet progress on this subject has been hindered by the lack of effective methods of video image segmentation. Lau et al report an objectbased video segmentation method where the segmentation was carried out manually by Lau et al., (2000). A recent Wavelet-based method for ultrasound image compression has been introduced by Chiu et al (2001). The basic ideas behind object-based video coding can be demonstrated using the examples in Figures 1 and 2. In Figure 1, a single video frame of an M-mode video is presented. Around the primary video image, there are four text-only regions labeled I, II, IV, and V in Figure 2, and outlined with extended white lines in Figure 1. We think of these video regions as video objects that will be independently encoded. In this case, it is clear that the difference between the text objects and the video objects will suggest that different Codecs should be used for encoding these rather distinct objects. For example, the text objects are very slow to change, while the video object changes substantially through time. Thus, it appears that more bandwidth must be allocated for the video images, as opposed to the text objects. On the other hand, it important to note that the lack of change in the text objects of the video will result in zero motion vectors, which in term will lead to less bandwidth allocated to these objects. Yet, if a special text encoder was used for text objects as opposed to the standard DCT encoder, we would expect significant improvement in text object compression efficiency. Thus, a key principle for achieving compression efficiency is to select an appropriate Codec for each object. In our research, we develop an object-based segmentation system based on a diagnostically lossless compression system. The basic idea is to segment out the part of the video image that is not of diagnostic significance. Then, by discarding the part of the video that is irrelevant to diagnosis, we can reduce the required bandwidth in the ratio of the discarded video to the entire video. The scheme is explained in detail in Section 4.

4. DIAGNOSTICALLY-LOSSLESS MOTION-MODE (M-MODE) VIDEO COMPRESSION We begin this Section with an introduction to the basic anatomical features of Motion-mode (M-mode) ultrasound video. We then proceed to define what we mean by diagnostically-lossless compression, as applied to compression of M-mode video.

OBJECT-BASED ULTRASOUND VIDEO

495

Near Field Anterior Right Ventricular Wall Right Ventricular chamber Interventricular septum Left Ventricular chamber with some mitral valve apparatus Left Ventricular posterior wall Bright epicardium

Figure 3. A basic Motion mode (M-mode) video image used in echocardiography.

The basic segmentation structure of a Motion-mode (M-mode) ultrasound video image used in echocardiography is shown in Figure 3. The basic principle in M-mode video is that we image the reflectivity of a single line of pixels through time. In Figure 3, as we move from the top to the bottom, we are moving through a spatial coordinate, from the point nearest the ultrasound probe to the point that is furthest away from the probe. From left to right we have the time evolution of the spatial line through time. Video images such as the one depicted in Figure 1 are typically used for assessing heart function. To recognize the video segments that are of diagnostic interest, each part of the image has been annotated. First, we have the near field that is of little diagnostic value. This is followed by the right ventricular wall, and the interventricular septum that separates the left from the right ventricle. The left ventricular chamber includes the mitral valve apparatus. The left ventricular chamber ends with the left ventricular posterior wall. This is followed by the epicardium. For assessing heart function, we are interested in clear images of the lower wall of the interventricular septum and the endocardial surface of the left ventricular posterior wall (LVPW).

4.1 Diagnostically-lossless Compression of M-mode Video: A Case Study To help define regions of diagnostic interest, an M-mode video image is segmented into rectangular regions as shown in Figure 4. These regions represent video objects that we have selected for diagnostically-lossless compression. In the following discussion, we discuss the diagnostic relevance of each region. We note that the near-field and the image portion beyond the epicardium is of no diagnostic interest. Thus, regions 1 and 3 are of no diagnostic value. The use of M-mode video in clinical diagnosis requires that we are able to clearly measure distances between wall boundaries. Thus, we require that the video compression algorithm does not degrade the definition of the wall boundaries. This is difficult to define in the sense that wall-boundaries themselves tend to be noisy, and in such cases,

P. RODRIGUEZ ET AL.

496

1. Near field Object. (extends up to 50th row)

2. Diagnostic Interest Object, (extends to just below 250th row).

3. Object beyond the epicardium.

Figure 4. Object-based segmentation of M-mode video. This video image was segmented using a frequency-modulation (FM) model.

sonographers are expected to extrapolate so that they get a well-defined, continuous line throughout the image.

For assessing the left ventricular function, we consider the shortening fraction, as defined by Snider at al. (page 196, 1997):

SF

LVDD  LVSD u 100 , LVDD

where: LVDD is the left-ventricular end-diastolic dimension LVSD is the left-ventricular end-systolic dimension SF is the shortening fraction (in percent). The shortening fraction has a normal mean value of 36% with an effective range between 28% and 44% , independent of age and heart rate. Thus, for evaluating the shortening fraction, it is not necessary to encode the walls during the entire cardiac cycle, but only during the time-instances when the posterior wall of the interventricular septum and anterior wall of the left ventricular wall come together (systole), or are furthest apart (diastole). Even though it is possible to achieve greater compression ratios by taking advantage of this observation, we will instead encode the entire wall region.

OBJECT-BASED ULTRASOUND VIDEO

497

This avoids the problem of having to develop a detector for systole and diastole, and also the problem of having to develop an encoder for curvilinear boundaries (shapeadaptive coding). Developing solutions for these problems is definitely possible. However, the complexity of a computer-assisted video compression system that incorporates these observations would increase substantially. This is due to the time that the sonographers would have to spend on verifying the detection of systole and diastole in every cycle, as opposed to simply verifying that the two object lines include the walls of interest.

5. M-MODE ULTRASOUND VIDEO OBJECT SEGMENTATION USING AMFM DEMODULATION The development of general image representations for object-based coding presents many challenges. There is a need for a spatio-temporal model that can describe (i) sharp edges that occur between objects, (ii) 3D object deformations or due to imaging of different slices through time, (iii) imaging artifacts such as ultrasound shadowing, and (iv) the presense of significant additive or multiplicative noise (specle noise in ultrasound). Edges are usually associated with very sharp transitions in image density, that traditional edge detectors try to capture by detecting large image gradients. For ultrasound images, there is a large number of such gradients associated with different reflecting boundaries. Furthermore, the boundaries turn to be broken, resulting in discontinuous edges that are difficult to link or track. Continuous-object deformations suggest that satisfactory edges can be achieved by developing methods that can assume continuity while removing the high levels of noise. We begin with a discussion of wellknown methods for noise reduction, and then continue with multi-resoultion and Wavelet based methods. We then present a novel Amplitude-Modulation Frequency-Modulation method that can be used to model edge continuity while providing for an effective methods for dealing with image noise. Traditional bandpass filtering methods are usually employed for reducing image noise by rejecting out-of-band components, which are assumed to be rich in noise, while containing less signal power. In practice, one starts by identifying the frequency bands where most of the signal power is present. The rest of the frequency bands are expected to be mostly due to noise and are thus rejected through bandpass filtering. The primary weakness of the approach is the assumption that signal components are made up of sinusoids that have infinite spatial extend, since the same bandpass is used throughout the image. Wavelet theory introduced multiresolution methods where images were modeled in terms of objects of finite extend, occurring at different resolutions. A basic assumption of Wavelet based filters is constant-Q filtering, where the ratio of the central frequency radius to the filter bandwidth. As a result, bandpass filters for higher frequency components occupy larger bandwidths, and correspondingly, they can be implemented using a smaller number of filter coefficients. It is important to note that these filters, with a reduced number of coefficients, also exhibit very high spatial resolution. The situation is reversed for low-frequency components. The bandpass filters have smaller bandwidths and exhibit very low spatial resolution. Thus, in order to describe the edges of larger

P. RODRIGUEZ ET AL.

498

objects, there is a need to combine different resolutions. This complicates the models since it is no longer possible to work at a fixed scale. The introduction of multidimensional Amplitude-Modulation Frequency-Modulation (AM-FM) methods attempts to overcome these limitations. Continuous-space and continuous-scale variations are modeled through the product of a spatially varying positive amplitude function and an FM signal. Throughout an object, the AM-FM model Step 1. Step 2. Step 3. Step 4.

Compute “Analytic Image” using 2-d FFT Apply bandpass channel filters Compute AM-FM parameters over each channel At each pixel, reconstruct AM-FM image using the channel that produces the maximum amplitude estimate.

Figure 5. An overview of an AM-FM demodulation system.

acts like the fundamental harmonic from Fourier Analysis. We assume slowly varying Amplitude Modulation (AM) and rely on Frequency Modulation to capture most intensity variation. The model is especially effective on images with a large number of ridges, such as Motion-mode ultrasound described in this Chapter. The fundamental advantage of the AM-FM representation is that it allows for a single model to describe the combined output from multiple bandpass filters. Both ridge orientation changes and variations in ridge spacing are captured through the Frequency Modulation model, while image intensity variations are captured in the amplitude. 5.1 An AM-FM Model for M-mode Ultrasound In this subsection, we introduce a video object segmentation system that is particularly well suited for M-mode video segmentation. To develop an AM-FM model for M-mode ultrasound, we return to Figure 1. In M-mode ultrasound video, the video is formed by plotting the beam reflection intensity along a line through time. The resulting video objects are made up of cardiac wall boundaries that evolve through time. There are significant challenges with modeling the walls, including broken boundaries, continuous-space deformations of the walls and speckle noise. To recognize a suitable AM-FM series for the M-mode ultrasound video image, we consider the row-column coordinate system y  t where y is used for indexing the rows, while the time coordinate t is used for indexing the columns. Along the y -coordinate, the ultrasound video image is "cutting through" cardiac wall boundaries, showing a single line of the 3-D object points. Along the time-coordinate t , it is assumed that the motion of the same 3-D wall points are observed through time. To track the wall motion, we consider the use of a curvilinear coordinate system that models wall-motion through a frequency modulation process, while variation in the actual wall material is described through an amplitude modulation process. We model M-mode video in terms of an AM-FM series expansion given by Rodriguez and Pattichis (2002), Pattichis (1998), and Havlicek (1996): f y, t ¦ C n a y, t cos nI y, t , n

where I denotes the phase function, cos nI y, t denotes the FM harmonics, a . denotes the amplitude function, and C n represents the coefficients of the AM-FM harmonics. We also define the instantaneous frequency as ’I , and its magnitude by the vector

OBJECT-BASED ULTRASOUND VIDEO

499

magnitude ’I . In what follows, we set x t to represent images in the familiar x  y coordinate system.

5.1 Real-time AM-FM Demodulation The basic AM-FM demodulation algorithm is summarized in Figure 5. Here, we define the AM-FM demodulation problem, as one of having to estimate the amplitude a x, y , the phase I x, y , and the instantaneous-frequency vector, as defined by Havlicek (1996):

’I x , y

>wI / wx, wI / wy @T ,

from the original image. To estimate the phase, we apply a collection of bandpass channel filters that cover the two-dimensional frequency plane at different orientations and magnitudes. The AMFM parameters are then estimated over each channel. If we let the impulse responses of the bandpass channel filters be denoted by g1 , g 2 , ..., g R , and let f denote the input

f * g i where image, the filtered output images are given by h1 , h2 , ..., hR satisfying hi * denotes two-dimensional convolution. We then obtain estimates for the instantaneous frequency and the phase using (see Havlicek (1996)): ­° imaginary ^hi ( x, y )` ½° ­ ’h x , y ½ ’Ii x, y # real® i ¾. ¾ , and Ii x, y # arctan ® ¯ jhi x, y ¿ ¯° real ^hi ( x, y )` ¿° Let G1 , G2 , ..., GR denote the frequency responses of the channel filters. Using the instantaneous frequency estimate ’I x, y and the frequency response of the channel

Gi , we estimate the amplitude over each channel using ai ( x, y ) #

hi x, y . Gi ’I x, y

In the Dominant Component Analysis (DCA) algorithm, from the estimates for each channel filter, we select the estimates from the channel with the maximum amplitude estimate cmax x, y argmax ai x, y . i

Hence, the algorithm adaptively selects the channel filter with the maximum response. This approach allows the model to quickly adapt to singularities in the image. The performance of the AM-FM demodulation algorithm in one and two-dimensions has been studied by Havlicek (1996). A discrete-space algorithm is also described by Havlicek (1996). Next, we present a number of modifications to the basic algorithm that allow us to achieve real-time performance. In addition, we discuss how to achieve robustness.

P. RODRIGUEZ ET AL.

500

5.2 Real-time AM-FM Demodulation Implementation In Step 1 of Figure 5, the 1D analytic signals are computed for each pixel-column in the image. We write:

X HX , i 0, 1, 2, ! , M  1. A, i i i where Xi denotes the i  th column in the image, M is the number of columns, and H denotes the Hilbert transform. X

The continuous-time analytic signal has a Fourier Transform that is equal to the Discrete Fourier Transform of the original signal for non-negative frequencies, and is zero for negative frequencies. This observation leads to a straight-forward implementation using 1D FFTs. For the bandpass filter implementation, we use a separable design. The separable implementation is efficient because the M-mode ultrasound video itself could be thought of as separable. This observation has been addressed in detail in Rodriguez and Pattichis (2002). In essence though, this is due to the fact that a product of vertical and horizontal modulations effectively capture the changes in the M-mode video.

Following AM-FM demodulation, a sum of 20 FM harmonics is taken as an approximation to the walls of the video image (see Rodriguez and Pattichis (2002)). The sum of the pixels along each row of the harmonics image is used for detecting the nearfield and the epicardium (see Figure 6). The sums from all the rows are filtered by a 19point median filter, followed by simple thresholding. This detection method was found to be very robust for segmentation (see Figure 6). 6. RESULTS

The rate distortion curves are summarized in Figure 9. We show the rate-distortion curves for: (i) method 1: single-object MPEG-4 video compression, (ii) method 2: MPEG-4 compression using four text objects and a video object, and (iii) method 3: MPEG-4 compression using FM-segmentation to derive a Region of Interest within the video object plus the four text objects. The video objects were encoded using the MPEG4IP software that is freely available (see MPEG4IP, 2003). In all Figures, the letters 1, 2, and 3 are used to identify the method that was used. Also, the dotted-line that overlaps with the rate-distortion curve of method 3 represents the result of multiplying the bitrate of method 2 by the ratio of the number of pixels in the FM-segmented image, divided by the number of pixels in the original image. It represents the theoretical performance improvement by using method 3.

OBJECT-BASED ULTRASOUND VIDEO

501

(a)

(b)

(c) Figure 6. M-mode video objects at different bitrates. In (a), both objects are shown at 50kbps, in (b) they are at 500kbps, and in (c) both objects are at 1000kbps.

P. RODRIGUEZ ET AL.

502

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(i)

Figure 7. M-mode video reconstructed with and without objects at different bitrates. In (a), (d), and (g) the entire video is compressed at 50Kbps, 500Kbps, and 1000Kbps. In (b), (c), (e), (f), (h), and (i), all the text objects are compressed at 10:1 using still image compression of the first frame. In (b), (e), and (h), we show the results of compressing the main video object at 50Kbps, 500Kbps, and 1000Kbps. Similarly, in (c), (f), and (i), we show the results of compressing the extracted region of diagnostic interest at 50Kbps, 500Kbps, and 1000Kbps.

OBJECT-BASED ULTRASOUND VIDEO

503

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

Figure 8. M-mode video object-based reconstruction. In (a), (b), (e) and (f), the effective bitrate is about 200Kbps. In (c), (d), (g), and (h), the effective bitrate is about 500Kbps.

P. RODRIGUEZ ET AL.

504

3

1

2 3

(a)

2

(b)

3

2

(c)

1

3

2

3 2

1

1

(d)

3

(e)

1

2

1

(f)

Figure 9. Rate-distortion curves using methods 1, 2, and 3. Method 1 is MPEG-4 without objects, method 2 is MPEG-4 with four text objects and one video object, while method 3 is MPEG-4 with four text objects and a segmented video object. The “leftmost” dotted curves show the result of multiplying the rate-distortion curve of method 2 by the ratio of the number of pixels in the segmented region divided by the entire video image region. It demonstrates that the FM segmentation method 3 achieves a near optimal improvement over method 2 that is proportional to the reduction in the number of video pixels used in method 3. Figures 9(a) and 9(b) correspond to the images in Figures 6 and 7. Figures 9(e) and 9(f) correspond to Figures 8(a)-8(d), while Figures 9(c) and 9(d) correspond to Figures 8(e)-8(h).

OBJECT-BASED ULTRASOUND VIDEO

505

For quantitative measurement of the compressed video quality, we compute the Psnr at a variety of bitrates. The Psnr is defined by § 2552 · Psnr 10 log10 ¨ ¸ © mse ¹

and measured in decibels (dB). In the formula for the Psnr, mse stands for the mean square error, defined as the average squared difference between the uncompressed video and the original (uncompressed) video intensity, while 255 represents the peak image intensity. For our study, Psnr values between 23dB and 25dB were found to be of interest (see Figure 9). A Psnr of 25dB represents a mean drop in image intensity of 14.34 (5.6% of 255), while a Psnr of 23dB represents a mean drop in image intensity of about 18.05, (7.1% of 255). Pnsr is not generally accepted as a good measure of perceptual image quality. Instead of using a Psnr measure over the entire image, we only use the Pnsr over the video region of diagnostic interest. As defined earlier, the Region of Interest (ROI) was estimated using FMsegmentation for extracting the M-mode video region that lies between the near field and the epicardium. Examples of the computation of the ROI are given in Figures 6, 7, and 8. In Figure 6, the rightmost images represent the ROIs. Furthermore, in the examples shown, all text objects were compressed using still-image compression of the first video frame, sacrificing the clock updates (which can be readily restored). To allow a critical evaluation of the rate-distortion curves, we also used perceptual evaluation of the compressed video. The purpose of the experiments was to determine the minimum bitrate below which the compressed video quality was unacceptable. The Cardiologist was allowed to review each video individually or compare videos against each other in order to determine whether the effect of compression was impacting the clinical use of the video. Two perceptual tests were carried out. In the first test, the reader was asked to assess the quality the displayed video. In the second test, the reader was asked to pause the video and to examine whether any distortion in the wall definition was likely to impact the diagnostic accuracy of the compressed video. The second test is in line with the standard practice of pausing the video playback before measurements on the video image are made. For both tests, three videos were tested at a variety of bitrates. For both tests, at high bitrates, 1000Kbps and above, no significant distortion was observed. To compare with other studies, we note that 500Kbps corresponds to a compression ratio of 50, 500Kbps corresponds to a compression ratio of about 120, and 200Kbps corresponds to a compression ratio of about 280. The fact that no significant distortion is observed at high bitrates is in agreement with the rate-distortion curves of Figure 9, where the Psnr appears to be flattened beyond 1000Kbps. As discussed earlier, at such low compression ratios, there is less than a 5% drop (from the peak value of 255) in image intensity between the compressed and the uncompressed videos. This very simple rule appears to be a good predictor of the diagnostically lossless region. For the first perceptual test, significant artifacts were observed in the neighborhood of 400Kps-500Kbps. This is in agreement with the sharp drop in the rate-distortion curves shown in Figure 9. For the second perceptual test, the minimum acceptable bitrate range varied from 100Kbps to 300Kbps. Thus, the cardiologist was able to detect artifacts in the continuous video display at a higher bitrate than in the still video image display. The continuous video display test is associated with a perceptually lossless requirement, while the still image video display

506

P. RODRIGUEZ ET AL.

test is associated with a diagnostically lossless requirement, where the basic question is whether there is sufficient information for a diagnosis. As expected, the perceptually lossless requirement was found to be stricter than the diagnostically lossless requirement. As seen in Figures 9(b), 9(d), and 9(f), over the diagnostically-relevant range, method 3 provided the same Psnr performance as both methods 1 and 2, but at a significant reduction of the bitrate (approximatedly 40% in all cases). The reduction in the bitrate is directly proportional to the reduction in the number of pixels used in the video object images of method 3 (see the rightmost images in Figures 6 and 7). At higher bitrates though, method 1 dominated. However, for wireless communications applications, there is obvious preference for the maximum acceptable compression achieved by method 3.

7. CONCLUDING REMARKS The Chapter has focused on the use of diagnostically-guided video compression. Based on clinical practice, video objects of interest were defined, and novel AM-FM segmentation methods were developed for extracting the objects. For each object, we examined quality (or SNR) scalability through rate-distortion curves. It is important to note the strong differences between the video objects in both coding requirements and their relevance to diagnosis. They range from text objects where only the clock text is changing, to near-field ultrasound video that is of no diagnostic value, to the main M-mode ultrasound video that includes the cardiac walls. It is also important to note that the near-field video, and the region beyond the epicardium would require high bandwidth for quality transmission. Yet, they are of no diagnostic interest. The different video objects require dramatically different types of Codecs. For the text objects, a text-based encoder that will first convert the text video objects into ascii followed by an alphabet based compression scheme such as LZW will produce optimal encoding. Furthermore, once the text has been converted to ascii, critical video content information associated with the acquisition can be extracted and used for digital libraries or other applications. For the m-mode video images, little bandwidth needs to be allocated to the near field or to the regions beyond the epicardium. In this case, most of the bandwidth will be allocated for tracking the cardiac wall motion. The coding gain is directly proportional to the number of pixels of the non-diagnostic objects divided by the total number of pixels. The difficult problem of developing quantitative measures that will be in agreement with perceptual experiments requires further investigation. However, over the region of diagnostic interest, compressed videos with mean intensity drop of less than 5% from the peak intensity value were found to be indistinguishable from the original (uncompressed) videos. This Chapter has focused on M-mode ultrasound. Further research is needed for establishing diagnostic measures and video objects of diagnostic interest for 2D-mode and the Doppler modes. It is also important to consider automated systems that can simultaneously, optimally compress video from combinations of different modes.

OBJECT-BASED ULTRASOUND VIDEO

507

8. REFERENCES Chiu, E., Vaisey, J., and Atkins, M. S., 2001, Wavelet-based space-frequency compression of ultrasound images, IEEE Transactions on Information Technology in Biomedicine., 5:300-310. Ehler, D., Vacek, J. L., Gowda, M., and Powers, K. B., 2000, Transition to an all-digital echocardiograpgy laboratory: a large, multi-site private cardiology practice experience, " Journal of the American Society of Echocardiography, 13:1109-1116. Garcia, M. J., Thomas, J. D., Greenberg, N., Sandelski, J., Herrera, C., Mudd, C., Wicks, J., Spencer, K., Neummann, A., Sankpal, B., and Soble, J., 2001, Comparison of mpeg-1 digital videotape with digitized svhs videotape for quantitative echocardiographic measurements, Journal of the American Society of Echocardiography, 14:114-121. ISO/IEC., 1999, IS 14496-X:Information technology-coding of audio-visual objects (MPEG-4). ISO/IEC., 1999, IS 14496-1:Information technology-coding of audio-visual objects-part 1: systems (MPEG-4 systems). ISO/IEC., 1999, IS 14496-2:Information technology-coding of audio-visual objects-part 2: visual (MPEG-4 video). Havlicek, J. P., 1996, AM-FM Image Models. Ph.D. diss., The University of Texas at Austin. Karson, T. H., Zepp, R. C., Chandra, S., Morehead, A., and Thomas, J. D., 1996, Digital storage of echocardiograms offers superior image quality to analog storage, even with 20:1 digital compression: results of the digital echo record access study, Journal of the American Society of Echocardiography, 9:769-778. Lau, C., Cabral, J. E., R. Jr, A. H., and Kim, Y., 2000, MPEG-4 Coding of Ultrasound Sequences, Proceedings of SPIE Medical Imaging 2000, 3976:573-579. Main, M. L., Foltz, D., Firstenberg, M. S., Bobinsky, E., Bailey, D., Frantz, B., Pleva, D., Baldizzi, M., Meyers, D. P., Jones, K., Spence, M. C., Freeman, K., Morehead, A., and Thomas, J. D., 2000, Real-time transmission of full-motion echocardiography over a high-speed data network: impact of data rate and network quality of service, Journal of the American Society of Echocardiography, 13:764-770. MPEG4IP, 2003; (2003) http://mpeg4ip.sourceforge.net/. Pattichis, M. S., 1998, AM-FM Transforms with Applications, Ph.D. diss., The University of Texas at Austin. Pattichis, M. S., Cai, S., Pattichis, C. S., and Abdallah, R., 2005, An overview of digital video processing algorithms, chapter in M-Health: Emerging Mobile Health Systems, R.H. Istepanian, S. Laxminarayan, and C.S. Pattichis, eds., Kluwer Academic/Plenum, New York. Rodriguez, P. V. and Pattichis, M.S., 2002, Real-time AM-FM Analysis of Ultrasound Video, invited in the 45th IEEE Midwest Symposium on Circuits and Systems 2002, Tulsa, Oklahoma, August 2002. Soble, J. S., Yurow, G., Brar, R., Stamos, T., Neumann, A., Garcia, M., Stoddard, M. F., Cherian, P. K., Bhamb, B., and Thomas, J. D., 1998, Comparison of MPEG digital video with super VHS tape for diagnostic echocardiographic readings, Journal of the American Society of Echocardiography, 11:819825. Snider, A. R., Serwer, G. A., and Ritter, S. B., 1997, Echocardiography in Pediatric Heart Disease. Second edition, St. Louis, Missouri, Mosby-Year. Thomas, J. D., Greenberg, N. L., and Garcia, M. J., Digital Echocardiography, 2002: Now is the time, Journal of the American Society of Echocardiography, 15:831-838.

object-based ultrasound video processing for wireless ... - Springer Link

The thin white lines define boundaries between objects of interest. In the upper rectangular region, patient information has been deleted. number of times, or if ...

794KB Sizes 1 Downloads 242 Views

Recommend Documents

Exploiting Graphics Processing Units for ... - Springer Link
Then we call the CUDA function. cudaMemcpy to ..... Processing Studies (AFIPS) Conference 30, 483–485. ... download.nvidia.com/compute/cuda/1 1/Website/.

Face Video Competition - Springer Link
web-cams and mobile devices embedded with a camera, it is now possible to re- ... Note that NIST MBGC aims at “portal application” where the task is to ... protocols reported in [4], when g1 is used as a development set (to build the user's.

Cross-layer modeling of adaptive wireless links for ... - Springer Link
May 8, 2006 - C Springer Science + Business Media, LLC 2006. Abstract Future wired-wireless multimedia networks re- quire diverse quality-of-service (QoS) ...

Are survival processing memory advantages based on ... - Springer Link
Feb 8, 2011 - specificity of ancestral priorities in survival-processing advantages in memory. Keywords Memory . ... have shown that survival processing enhances memory performance as measured by recall, ..... Nairne, J. S., Pandeirada, J. N. S., & T

Auditory motion processing in early blind subjects - Springer Link
Although the neural substrates of visual motion pro- cessing have been extensively researched for several decades (for a review, see Culham et al. 2001), little is known about auditory motion processing. The few neuroimaging studies investigating aud

How inherently noisy is human sensory processing? - Springer Link
Like any physical information processing device, the human brain is inherently noisy: If a participant is presented with the ... mental and computational efforts aimed at understanding and characterizing human sensory processing. Psychonomic ..... we

Are survival processing memory advantages based on ... - Springer Link
Published online: 8 February 2011. © Psychonomic Society, Inc. 2011 ... study investigated the specificity of this idea by comparing an ancestor-consistent ...

Contributions of beliefs and processing fluency to the ... - Springer Link
Nov 27, 2012 - Abstract Discovering how people judge their memories has been a major issue for metacognitive research for over. 4 decades; many factors ...

Evidence for Cyclic Spell-Out - Springer Link
Jul 23, 2001 - embedding C0 as argued in Section 2.1, this allows one to test whether object ... descriptively head-final languages but also dominantly head-initial lan- ..... The Phonology-Syntax Connection, University of Chicago Press,.

MAJORIZATION AND ADDITIVITY FOR MULTIMODE ... - Springer Link
where 〈z|ρ|z〉 is the Husimi function, |z〉 are the Glauber coherent vectors, .... Let Φ be a Gaussian gauge-covariant channel and f be a concave function on [0, 1].

Tinospora crispa - Springer Link
naturally free from side effects are still in use by diabetic patients, especially in Third .... For the perifusion studies, data from rat islets are presented as mean absolute .... treated animals showed signs of recovery in body weight gains, reach

Chloraea alpina - Springer Link
Many floral characters influence not only pollen receipt and seed set but also pollen export and the number of seeds sired in the .... inserted by natural agents were not included in the final data set. Data were analysed with a ..... Ashman, T.L. an

GOODMAN'S - Springer Link
relation (evidential support) in “grue” contexts, not a logical relation (the ...... Fitelson, B.: The paradox of confirmation, Philosophy Compass, in B. Weatherson.

Bubo bubo - Springer Link
a local spatial-scale analysis. Joaquın Ortego Æ Pedro J. Cordero. Received: 16 March 2009 / Accepted: 17 August 2009 / Published online: 4 September 2009. Ó Springer Science+Business Media B.V. 2009. Abstract Knowledge of the factors influencing

Quantum Programming - Springer Link
Abstract. In this paper a programming language, qGCL, is presented for the expression of quantum algorithms. It contains the features re- quired to program a 'universal' quantum computer (including initiali- sation and observation), has a formal sema

BMC Bioinformatics - Springer Link
Apr 11, 2008 - Abstract. Background: This paper describes the design of an event ontology being developed for application in the machine understanding of infectious disease-related events reported in natural language text. This event ontology is desi

Isoperimetric inequalities for submanifolds with ... - Springer Link
Jul 23, 2011 - if ωn is the volume of a unit ball in Rn, then. nnωnVol(D)n−1 ≤ Vol(∂D)n and equality holds if and only if D is a ball. As an extension of the above classical isoperimetric inequality, it is conjectured that any n-dimensional c

Probabilities for new theories - Springer Link
where between 0 and r, where r is the prior probability that none of the existing theories is ..... theorist's internal programming language"(Dorling 1991, p. 199).

A Process Semantics for BPMN - Springer Link
Business Process Modelling Notation (BPMN), developed by the Business ..... In this paper we call both sequence flows and exception flows 'transitions'; states are linked ...... International Conference on Integrated Formal Methods, pp. 77–96 ...

Unsupervised Learning for Graph Matching - Springer Link
Apr 14, 2011 - Springer Science+Business Media, LLC 2011. Abstract Graph .... tion as an integer quadratic program (Leordeanu and Hebert. 2006; Cour and Shi ... computer vision applications such as: discovering texture regularity (Hays et al. .... fo

Candidate quality - Springer Link
didate quality when the campaigning costs are sufficiently high. Keywords Politicians' competence . Career concerns . Campaigning costs . Rewards for elected ...

Mathematical Biology - Springer Link
Here φ is the general form of free energy density. ... surfaces. γ is the edge energy density on the boundary. ..... According to the conventional Green theorem.