MACHINE LEARNING BASED MODELING OF SPATIAL AND TEMPORAL FACTORS FOR VIDEO QUALITY ASSESSMENT Manish Narwaria and Weisi Lin School of Computer Engineering, Nanyang Technological University, Singapore ABSTRACT Unlike image quality, video quality is affected by the temporal factor, in addition to the spatial one. In this paper, we investigate into the impact of both the factors on the overall perceived video quality and combine them into a metric. We use machine learning as a tool to study and analyze the relationship between the factors and the overall perceived video quality. It is shown that apart from their individual contributions, the interaction of the two factors also plays a role in determining the overall video quality. We report the experimental results and the related analysis using videos from two publicly available databases. Index Terms— Video quality assessment (VQA), machine learning, spatial quality, temporal quality 1. INTRODUCTION Objective video quality assessment (VQA) has emerged as an important research in numerous multimedia applications. Owing to the limitations of subjective assessment (such as high cost, unsuitability for many real-time applications etc.), objective VQA to predict subjective viewing results has attracted significant attention during the recent years [1][10]. There are two broad issues involved in objective VQA. First, given a test video, the spatial and temporal scores need to be determined. Secondly, the two scores should be combined to obtain an overall video quality score. Survey of literature shows that the first issue has received more research attention as is evident by many recent works [1][10] which attempt to measure quality along both spatial and temporal dimensions in different ways. However, the second issue regarding the combination of the two factors is relatively uninvestigated. Some of the existing VQA metrics simply use temporal factors as the weighting (for e.g. [2], [5], [10]) for the spatial factor while others (e.g. [1], [4], [6], [9]) combine the two via a pre-defined relationship using parameters. However, such approaches are not convincing and less effective due to being ad-hoc. Furthermore, the methods of selecting the parameter values are also less satisfactory (e.g. empirical selection of parameter values [3],

[9]). We believe that for effective quality prediction the contribution of spatial and temporal factors as well as their interaction needs to be combined more intelligently. One solution is to use machine learning approaches as they are expected to be more convincing and meaningful due to being data-driven. Such approaches have been relatively unexplored in the context of modeling of the video quality. In our opinion, they will be more reasonable and powerful for feature combination in comparison to the existing ad-hoc methods which generally use a trial and error approach. Therefore, in this paper, we exploit machine learning for the said task and demonstrate its effectiveness and generality by way of extensive experimental analysis. 2. THE PROPOSED METHOD In this section, we describe the proposed method to predict video quality by combining the spatial and temporal factors through a relationship derived via training with ground truth (i.e. subjective scores). To demonstrate the general applicability of the proposed idea, we use two independent VQA metrics proposed in [2] and [9] for the calculation of spatial and temporal factors. So we first briefly introduce these two metrics and then propose the new metric in this work. 2.1 Motion-based Video Integrity Evaluation index The first metric used in this study is the Motion-based Video Integrity Evaluation index (MOVIE). It uses Gabor filter to decompose the reference and distorted videos into spatiotemporal bandpass channels. Motion information is computed from the reference video sequence in optical flow fields. The set of Gabor filters used to compute the spatial quality is also used to calculate optical flow from the reference video. The reader is referred to [2] for further details. The spatial and temporal scores are denoted as Smovie and Tmovie respectively and the overall quality MOVIE has been defined [2] as a product MOVIE = Smovie x Tmovie (1) We used the default values of the parameters given in [2] to compute Smovie and Tmovie.

2.2 Temporal speed similarity index (TSSI) based metric The second metric that we consider [9] uses multi-scale SSIM (MS-SSIM) [11] and motion vector similarity to compute the spatial and temporal scores respectively. Consider a video sequence with F frames. Let SQi denote the spatial quality of the ith frame using MS-SSIM. Then video sequence’s spatial quality score S is obtained by using a simple time average as S =

1 F

F

∑ SQ

i

(2)

i =1

A higher S indicates better spatial quality. The temporal score is computed by measuring the similarity between the motion vectors (MVs) of the reference and the distorted video sequences. The temporal speed similarity index (TSSI) is then defined as

2vr vd + C vr2 + vd2 + C

TSSI (ref , dis) =

(3)

where vr and vd are the speeds of the motion (magnitudes of the MVs) for the blocks ref and dis in the reference and distorted video frames respectively. A block-based motion estimation algorithm was used to estimate the MVs. The constant C (=√2 x search size) in (3) was used to avoid instability when the denominator is close to zero. For a video with mn blocks in each frame, the mean temporal score T is given as 1 (4) T = TSSI Fmn F m n

∑∑∑

The expression in (3) computes the similarity between the MVs of the reference and distorted video such that 0 ≤ TSSI ≤ 1. TSSI = 1 for a perfect quality video while its value will be lesser for a lower quality video. It follows that a higher T indicates better temporal quality. The authors of [9] combined S and T to obtain the overall quality score Q using two parameters as follows p2

 S  (5) Q= p  T  where p1 = 0.37 and p2 = 10 were determined empirically. 1

As will be discussed in Section 3.2, (5) has defects for its purpose. 2.3 Adaptive Basis Function Regression In this paper, we use the adaptive basis function regression (ABFR) [12] for combining the spatial and temporal scores. It is a modified form of the polynomial regression in which the basis functions are determined adaptively from the training data and is thus more effective. Let d be the number of input variables, r be a k×d matrix of nonnegative integer exponents such that rij is the exponent of the jth variable in the ith basis function. Note that, when for a particular ith basis

function rij = 0 for all j, the basis function is the intercept term. The matrix r completely defines the structure of the polynomial model with all its basis functions. The set of basis functions is then defined as  f = 

d

∏x j =1

rij j

 | i = 1, 2,...k  

(6)

Formally, the problem of finding the best set of basis functions can be defined as finding the best matrix r with the best combination of nonnegative integer values of its elements  r * = arg min J  r 

d

∏x j =1

rij j

  

(7)

where J(.) is an evaluation criterion that evaluates the predictive performance of the regression model corresponding to a set of basis functions. In this paper, we used the Corrected Akaike’s Information Criterion [14]. The reader is referred to [12] for further details regarding ABFR. We used the spatial and temporal scores as the input with the subjective video quality score being the target value. Therefore, we have a 2-dimensional input (i.e. d = 2) vector x = {x1 , x 2 } where x1 = spatial score and x2 = temporal score. 3. EXPERIMENTAL RESULTS AND ANALYSIS 3.1. Database and Test methodology We have used the LIVE video database [7]. In this database, 15 distorted videos have been generated for each reference video by using four distortion processes namely simulated transmission of H.264 compressed bit streams through errorprone wireless networks, simulated transmission through error-prone IP networks, H.264 compression and MPEG-2 compression. The subjective scores have been made available as difference mean opinion scores (DMOS). The authors of [9] used only the first 7 reference videos (out of a total of 10) and their distorted versions, i.e., 105 distorted videos. Since we obtained the S and T scores directly from the authors of [9] for convincing comparison, we used the 105 S and T scores provided by them and the corresponding subjective scores for these videos. Similarly, Smovie and Tmovie were computed for these 105 videos using the software [7] provided by the authors of MOVIE. We used the scores from 4 reference videos as the training set (denoted as TrainingLIVE) while the test set consisted of scores from the remaining 3 videos. Therefore, the training and testing sets come from disjoint sets since videos used for training are not used for testing. The training and test sets consisted of 60 and 45 datapoints respectively. A 4parameter monotonic logistic mapping between the objective outputs and the subjective quality ratings was also

employed, following the Video Quality Experts Group (VQEG) Phase-I/II test and validation method [8]. The experimental results are reported in terms of four criteria namely: Pearson linear correlation coefficient CP (for prediction accuracy), Spearman correlation coefficient CS (for monotonicity), Kendall rank correlation coefficient CK and the Root Mean Squared Error (RMSE), between the subjective score and the objective prediction. A better quality metric has higher CP, CK, CS and lower RMSE. 3.2 Analysis with ABFR Based on training, we obtained the following relationship between the overall video quality QABFR and the individual factors S and T Q ABFR = 134 − 28.2 S 3 − 62.1 S T 2 (8) This equation reveals that the basis functions selected by ABFR based on training are

{

3

2

}

f =1, S , S T The above set of basis functions shows that apart from the individual contribution from S, the interaction (denoted by the multiplicative term S T 2 ) term also plays a role in quality prediction. This is in agreement with observations in many previous works (e.g. [1], [6], [10], [16]) that the human perception is also affected by the interaction between the spatial and temporal factors. An explanation for the importance of the interaction term is that it can be thought as the overlap between the two factors and therefore represents the adjustment for the combined effect in perception [13]. As mentioned before, many existing works use only the interaction term (or a modified form such as Equation (5)) as the video quality measure. However, as suggested by Equation (8), the relationship between different factors and overall quality can be more complex and non-linear. The experimental results for QABFR are presented in Table 1. We find that it performs better than Q and S. We believe that Q defined in Equation (5) lacks justification in terms of the parameters and the functional form chosen. As explained in Section. 2.2, from a qualitative point of view, both S and T will be bigger for a better quality video and smaller for lower quality video. Since they follow the same trend of increase or decrease, a product of S and T (or their sum) would be a more intuitive choice for the overall quality measure. However, as seen in Equation (5), they are combined via an inverse relationship (using two parameters) which does not agree intuitively. Interestingly, as seen in Table 1, S performs better than Q for the 45 test videos in spite of Q using the temporal factor T. In summary, although ABFR leads to a small performance improvement, its use results in a more convincing and intuitive relationship. Next we used Smovie and Tmovie and obtained the following relationship for the overall video quality MOVIEABFR:

2 MOVIE ABFR = 11 .4 + 9.75Tmovie + 4.59 S movie − 0.97Tmovie 2 3 +0.04 S movie Tmovie + 0.02Tmovie + 0.36 S movie Tmovie

(9)

Therefore the basis functions are 2 2 3 f = 1 , Tmovie , Smovie , Tmovie , Smovie Tmovie , Tmovie , Smovie Tmovie

{

}

As seen in Equation (9), the last term is the same as Equation (1) which suggests that the metric MOVIE can be considered as a special case of MOVIEABFR. That is the contribution from other terms apart from the interaction term that should also be considered. The experimental results for MOVIEABFR are presented in Table 2. We can see that it performs better than MOVIE which in turn performs better than Smovie. This is expected since the improvement in the performance of MOVIE as compared to Smovie is due to the use of the temporal factor. On the other hand, MOVIEABFR derived via training is more effective and general. Since the software for MOVIE is available, we conducted additional experiments for it using another publicly available video database (we refer to it as EPFL database) [15]. It consists of 78 video streams (along with their subjective scores) encoded with H.264/AVC and corrupted by simulating the transmission over error-prone network. The related theoretical analysis and discussion is similar to that of Equation (9), so we include only the experimental results. We considered training the system in two different ways. First, we partitioned the data such that 39 videos are used for training while the remaining 39 videos form the test set. This ' . Secondly, we used cross test case is denoted as MOVIE ABFR

database validation, i.e., the same test set of 39 videos was tested with the training set being TrainingLIVE (defined in Section 3.1). This test case is denoted by MOVIEABFR(LIVE) (which means training is done with LIVE database). The experimental results are presented in Table 3. It is interesting to note that overall Smovie and MOVIE give quite close performance in this case. This is in contrast to the results obtained for the LIVE video database (see Table 2) where MOVIE performed better than Smovie. This indicates that the incorporation of the temporal factor via only the interaction term (as in Equation (1)) is less effective. On the ' other hand, as expected, MOVIE ABFR performs relatively better for all 4 test criteria. In addition, we find that MOVIEABFR(LIVE) also performs quite well which is significant since the training and test sets come from different video databases.

3.3 Further Discussion In previous works [1], [16], subjective experiments were used to determine that interaction term is important. By contrast, we have used machine learning to show that interaction term plays a role. We have further argued that the

Table 1. Performance comparison for the metric in [9] on LIVE video database Measure S Q QABFR CP 0.8005 0.7990 0.8115 CS 0.7689 0.7713 0.7734 CK 0.5697 0.5677 0.5798 RMSE 6.0066 6.0807 5.9097 Table 2. Performance comparison for the metric in [2] on LIVE video database Measure Smovie MOVIE MOVIEABFR CP 0.6234 0.6633 0.7112 CS 0.6032 0.6318 0.6860 CK 0.4124 0.4586 0.5070 RMSE 7.0150 6.7138 6.3070 Table 3. Performance comparison for the metric in [2] on EPFL video database MOVIE MOVIE’ABFR MOVIEABFR(LIVE) Measure Smovie CP 0.9230 0.9228 0.9302 0.9244 CS 0.8645 0.8656 0.8842 0.8650 CK 0.6702 0.6772 0.6923 0.6950 RMSE 0.5350 0.5356 0.5219 0.5300

interaction term alone may not be sufficient and used datadriven approach to arrive at a more reasonable model for the overall quality prediction. Since it is not easy to know the contribution of each factor apriori, we believe it is better to employ machine learning to determine the related parameters via training. Additionally, the basis functions obtained through ABFR provide more insights about the established relationship. As mentioned before, the training and test sets were chosen such that there is no overlap between the two with regards to video contents. Besides, the good performance for the cross-database validation test further indicates the system’s robustness to untrained video and/or distortion contents.

factors for overall quality determination. ACKNOWLEDGEMENT The authors wish to thank T. Liu (Arizona State University) for providing the spatial and temporal scores and Dr. K. Seshadrinathan for giving the software for MOVIE. 5. REFERENCES [1] Q. Thu and M. Ghanbari, “Modelling of spatio–temporal interaction [2] [3]

[4] [5] [6] [7] [8] [9] [10]

[11]

4. CONCLUSION While much of the research effort in VQA has concentrated on the analysis and computation of spatial and temporal factors, the issue of combining the factors has largely remained uninvestigated. In this paper, we explored the use of machine learning for a proper combination of different factors towards more effective VQA. We carried out analysis and experiments using the spatial and temporal factors of two third party VQA metrics and two publicly available video databases. The results in this study suggest that non-linear combination of factors can help in improving the prediction performance. An important contribution of this work has been the use of machine learning to obtain the relationship for video quality prediction. The results of this study also provide new insight for video quality evaluation regarding the issue of combining spatial and temporal

[12] [13] [14] [15]

[16]

for video quality assessment”, Signal Processing: Image Communication, Vol. 25, 2010, pp. 535-546. K. Seshadrinathan and A. Bovik, “Motion Tuned Spatiotemporal Quality Assessment of Natural Videos”, IEEE Trans. on Image Processing, Vol. 9, no. 2, 2010. A. Ninassi, O. L. Meur, P. Callet, and D. Barba, “Considering temporal variations of spatial visual distortions in video quality assessment,” IEEE Journal of Selected Topics in Signal Processing, vol. 3, no. 2, 2009, pp. 253–265. M. Barkowsky, B. Bialkowski, R. Bitto, and A. Kaup, “Temporal trajectory aware video quality measure,” IEEE Journal of Selected Topics in Signal Processing, vol. 3, no. 2, pp. 266–279, April 2009. Z. Wang and Q. Li, “Video quality assessment using a statistical model of human visual speed perception.” Journal of Optical Society of America, vol. 24, no. 12, 2007, pp. B61–B69. M.H. Pinson, S. Wolf, “A new standardized method for objectively measuring video quality”, IEEE Transactions on Broadcasting, Vol. 50, no. 3, 2004, pp. 312–322. LIVE Video Quality Database, 2009. [Online]. Available: http://live.ece.utexas.edu/research/quality/live_video.html. VQEG, Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment, Phase II August 2003 [Online]. Available: http://www.vqeg.org. T. Liu, K. Liu and H. Liu, “Temporal information assisted video quality metric for multimedia”, Proc. of IEEE International Conference on Multimedia and Expo (ICME), 2010, pp. 697 – 701. R. Vidal and J. Gicquel, “A No-Reference Video Quality Metric based on a Human Assessment Model”, in Proc. International Conference on Video Processing and Quality Metrics for Consumer Electronics (VPQM 07), 2007. Z. Wang, E. Simoncelli, and A. Bovik, “Multiscale structural similarity for image quality assessment,” in IEEE Asilomar Conference on Signals, Systems and Computers, 2003, pp. 1398 1402. G. Jekabsons, “Adaptive Basis Function Construction: an approach for adaptive building of sparse polynomial regression models”, Machine Learning, Yagang Zhang (ed.), 2010, pp. 127-156. H. Nothdurft, “Salience from feature contrast: additivity across dimensions”, Vis. Res., vol. 40, no. 10–12, 2000, pp. 1183–1201. C. Hurvich and C. Tsai, “Regression and time series model selection in small samples”, Biometrika, Vol. 76, 1989, pp. 297-307. F. Simone, M. Naccari, M. Tagliasacchi, F. C. Dufaux, S. Tubaro and T. Ebrahimi, “Subjective Assessment of H.264/Avc Video Sequences Transmitted Over A Noisy Channel", Proc. of IEEE International Workshop on Quality of Multimedia Experience 2009, pp. 204-209. C. Mantel, T. Kunlin and P. Ladret, “The Role of Temporal Aspects for Quality Assessment”, Proc. of IEEE International Workshop on Quality of Multimedia Experience 2009, pp. 94-99.

MACHINE LEARNING BASED MODELING OF ...

function rij = 0 for all j, the basis function is the intercept term. The matrix r completely defines the structure of the polynomial model with all its basis functions.

157KB Sizes 1 Downloads 207 Views

Recommend Documents

BASED MACHINE LEARNING METHODS FOR ...
machine learning methods for interpreting neural activity in order to predict .... describing the problems faced in motor control and what characterizes natural.

MALADY: A Machine Learning-based Autonomous ...
One of the characteristics that distinguishes sensor networks ... MALADY, that can be used by the nodes in a sensor network ... wireless sensor networks.

MALADY: A Machine Learning-based Autonomous ... - Semantic Scholar
Geethapriya Thamilarasu. Dept. of Computer Science & Engineering ... when there is some degree of imprecision in the data. Third, in some cases, example ...

MALADY: A Machine Learning-based Autonomous ... - Semantic Scholar
MALADY: A Machine Learning-based Autonomous ... machine learning techniques for autonomous decision making. .... on MicaZ motes running TinyOS.

Machine learning-based 3D resist model
94.5 signals. We performed 10-fold cross-validation of two alternative ANNs. In case of ANN for regression,. ANN in turn had 69 input nodes and 5 hidden layers, ...

Machine learning-based 3D resist model
Accurate prediction of resist profile has become more important as ... ANN corresponds to our machine learning-based resist 3D model (ML-R3D model). Due to ...

Applied Machine Learning - GitHub
In Azure ML Studio, on the Notebooks tab, open the TimeSeries notebook you uploaded ... 9. Save and run the experiment, and visualize the output of the Select ...

Machine learning - Royal Society
a vast number of examples, which machine learning .... for businesses about, for example, the value of machine ...... phone apps, but also used to automatically.

Applied Machine Learning - GitHub
Then in the Upload a new notebook dialog box, browse to select the notebook .... 9. On the browser tab containing the dashboard page for your Azure ML web ...

Machine learning - Royal Society
used on social media; voice recognition systems .... 10. MACHINE LEARNING: THE POWER AND PROMISE OF COMPUTERS THAT LEARN BY EXAMPLE ..... which show you websites or advertisements based on your web browsing habits'.

Applied Machine Learning - GitHub
course. Exploring Spatial Data. In this exercise, you will explore the Meuse ... folder where you extracted the lab files on your local computer. ... When you have completed all of the coding tasks in the notebook, save your changes and then.

Gaussian Margin Machines - Proceedings of Machine Learning ...
separable samples, we can relax the inequality constraints by introducing a slack variable ξi for each point xi and aug- menting the objective function with a ...

Overview of Machine Learning and H2O.ai - GitHub
Gradient Boosting Machine: Highly tunable tree-boosting ensembles. •. Deep neural networks: Multi-layer feed-forward neural networks for standard data mining tasks. •. Convolutional neural networks: Sophisticated architectures for pattern recogni

Exchangeable Variable Models - Proceedings of Machine Learning ...
Illustration of low tree-width models exploiting in- dependence (a)-(c) and .... to the mixing weights wt; then draw three consecutive balls from the chosen urn ..... value to 1 if the original feature value was greater than 50, and to 0 otherwise.

Machine Learning In Chemoinformatics - International Journal of ...
Support vector machine is one of the emerging m/c learning tool which is used in QSAR study ... A more recent use of SVM is in ranking of chemical structure [4].

Deep Boosting - Proceedings of Machine Learning Research
We give new data-dependent learning bounds for convex ensembles. These guarantees are expressed in terms of the Rademacher complexities of the sub-families. Hk and the mixture weight assigned to each Hk, in ad- dition to the familiar margin terms and

Machine Learning of User Profiles: Representational Issues
tools for finding information of interest to users becomes increasingly ... Work on the application of machine learning techniques for constructing .... improved retrieval performance on TIPSTER queries, and to further ... testing procedure.

Gaussian Margin Machines - Proceedings of Machine Learning ...
we maintain a distribution over alternative weight vectors, rather than committing to ..... We implemented in matlab a Hildreth-like algorithm (Cen- sor and Zenios ...

Batch Normalization - Proceedings of Machine Learning Research
2010) ReLU(x) = max(x, 0), careful initialization (Ben- gio & Glorot, 2010; Saxe et al., 2013), and small learning rates. If, however, we could ensure that the distribution of nonlinearity inputs remains more stable as the network trains, then the op