User Evaluation of an Interactive Music Information Retrieval System Xiao Hu

Noriko Kando

Xiaojun Yuan

Library and Information Science University of Denver 1999 E. Evans Ave. Denver, CO, 80208, U.S.A.

National Institute of Informatics 2-1-2 Hitotsubashi, Chiyoda-ku, Tokyo 101-8430 Japan

College of Computing and Information University at Albany, State University of New York, 135 Western Avenue Albany, NY 12222

[email protected]

[email protected]

[email protected]

ABSTRACT Music information retrieval (MIR) is a highly interactive process and MIR systems should support users’ interactions during the process. However, there are very few user-centered experiments in evaluating interactive MIR systems. In this paper, we present a user experiment with 32 participants evaluating two result presentation modes of a music mood recommendation system for two music seeking tasks, using user-centered measures. Preliminary results disclosed that the presentation modes made significant difference on some of the user-centered measures and users may enjoy the less effective but more appealing mode.

to the rapid growth of MIR, but there has not been an interactive or user-centered task since its inception in 2004.

evaluation, music information retrieval, user-centered evaluation, music mood recommendation, result presentation.

As one of the first efforts to fill this gap, this study takes a usercentered approach to evaluate two result presentation modes of an end-user MIR system on two music seeking tasks, namely searching and browsing. In particular, we focus on seeking music by mood, because mood is one of the most important reasons for people’s engagement with music [5], and recent user studies in MIR have identify mood as a major access point to music (e.g., [7][10]). The system being evaluated in this study is moodydb, an interactive music mood recommendation system helping users find music pieces in certain mood [4]. We chose this system for two reasons. First, it used the same classification techniques of the best performing systems in the Audio Mood Classification (AMC) task in MIREX 2007 [3] whose performances are still among the best to date. 2) It provides two modes of presenting music recommendations: one is the conventional list-based layout (called “list”); the other is a novel size-based layout (called “visual”). User-centered comparison of the two modes can shed light on the impact of interactive interfaces on music information seeking process. The research question we strive to answer in this study is whether the list and visual presentation modes have different influences on user-centered measures for the tasks of finding songs in certain mood.

1. INTRODUCTION

2. THE SYSTEM

The process of seeking music information is inherently interactive, as music appreciation is a highly subjective and fluid experience. In many cases, users cannot clearly describe their information needs when they set out to search for music. During a music search, it is not uncommon that users may find their needs more or less changed. In order to help users search music more efficiently, it is important for MIR systems to support interactive and exploratory search.

Moodydb is a content-based music mood classification and retrieval system. It is accessible online using any Web browser. It extracts salient spectral features (e.g., Mel-frequency Cepstral Coefficients) from music audio and classifies music pieces into five mood categories: passionate, cheerful, bittersweet, silly/quirky and aggressive. The five categories are borrowed from the Audio Mood Classification (AMC) task in MIREX [3]. The classification model was built by the SMO (sequential minimal optimization) implementation of Support Vector Machines in the Weka toolkit [11]. For each music piece, moodydb calculates the probability of this piece belonging to each of the five mood categories and builds a mood profile for the piece using a five dimensional vector with each dimension being one of the five probabilities. The similarity between two music pieces is then calculated by the distance between their mood profiles.

Categories and Subject Descriptors H.3.3 [Information Storage and Retrieval]: Information Search and Retrieval – search process, H.5.1 [Information Interfaces and Presentation]: Multimedia Information Systems – Evaluation/Methodology, J.5 [Art and Humanities] – Music

General Terms Measurement, Performance, Human Factors.

Keywords

In the last decade, many Music Information Retrieval (MIR) systems have been developed to serve users music information needs. Interface of a MIR system is the bridge between users and system functions. To date, most MIR systems and online services have been using the list-based layout to present search results or recommendations, which was adopted directly from the domain of text information retrieval. MIR researchers have been developing alternative presentation methods such as album grid [6] and selforganized maps [8]. However, very few formal user experiments have been conducted to evaluate MIR interfaces using useroriented approaches. In fact, MIR evaluation has been overwhelmingly dominated by system-centered approaches. The most influential evaluation event, the Music Information Retrieval Evaluation eXchange (MIREX) [1], has made great contribution

When a user starts using moodydb, he or she types in part of a song title or artist name. Moodydb will display a list of songs matching the textual query. After the user selects one of the songs as the seed, moodydb will retrieve a set of songs with similar mood profiles to the seed song, and display them on the screen. The user then can examine and play the songs. For each song, basic metadata (e.g., title, artist) and the image of album cover are

displayed. Due to intellectual property constraints, each audio clip in moodydb is 30 seconds long extracted from the middle of a song. Moodydb provides two modes of presenting recommended songs, named list and visual. Figure 1 shows the list mode with important areas marked. The list mode ranks recommended songs from top to bottom of the page according to their mood similarity to the seed song. The higher a song is listed, the more similar it is to the seed song. The visual mode is shown in Figure 2. It uses the sizes of album images to indicate the similarity levels. The larger a song’s album image is, the more similar it is to the seed song. The list mode can display up to 10 recommended songs per page while the visual mode can display 15 per page.

3.2 Tasks Inspired by real-life music seeking experiences, two user tasks were defined for the experiment, searching and browsing. In a searching task, a user is given one seed song, and the goal is to find other songs with similar mood to the seed song. In a browsing task, a user chooses one of the mood categories he or she is interested in, and the goal is to discover songs he or she likes. A searching task is similar to the “query-by-example” scenario fairly common in MIR, but the criterion here emphasizes mood similarity. A browsing task is less strictly defined, simulating the exploratory research scenario where users’ goals are not very well defined. For each of the tasks, two specific topics were designed. The two topics of searching task differed in the seed songs given to the users. For the browsing task, users were given the title and artist of an exemplar song for each of the mood categories, and the two browsing topics had different sets of exemplar songs. Table 1. Distribution of songs in moodydb. Mood passionate cheerful bittersweet silly/quirk aggressive Total

Western songs 184 164 215 35 140 738

Chinese songs 0 0 11 0 1 12

Total 184 164 226 35 141 750

3.3 Participants Figure 1. The list result presentation mode in moodydb.

32 Japanese undergraduate and graduate students from 13 different universities were recruited, including 14 females and 18 males. Their average age was 21.1 years old (standard deviation was 1.67), and their majors ranged from engineering, medicine to social sciences and humanities. Statistics of participants’ background on music knowledge, music listening, searching, computer and English skills are shown in Table 2 and Figure 3. Self-reported English abilities were collected because most of the songs had English lyrics and the answer sheet handouts were written in English (see below). Table 2. Statistics of participants’ background (each item was measured by a Likert scale from 1 to 7. 1: novice, 7: expert). Music knowledge Expertise with computers Expertise with online searching Ability in reading English Ability in listening to English

Figure 2. The visual result presentation mode in moodydb. There are 750 songs in moodydb. Most of them are Western popular music (from U.S. and U.K.). Table 1 shows the distribution of the songs in terms of mood categories and sources.

3. METHOD 3.1 Experimental Design In the experiment, we focused on the influences of the result presentation modes (list and visual) and tasks (searching and browsing). The participants were assigned to a factorial experiment that included two modes, two tasks and two topics per task as within-subject factors. As explained in details below, these factors and their orders were counter-balanced.

Mean 3.81 4.25 4.44 4.91 4.22

Standard deviation 1.12 0.88 1.01 0.82 1.13

3.4 Procedure The experiment was conducted in a batch manner, with 4 participants in each batch performing the tasks at the same time. There were 8 batches in total. Participants filled a pre-experiment questionnaire at the beginning and a post-experiment questionnaire at the end. Before conducting the tasks, there was a 10 minutes training sessions on how to use the moodydb system. Each participant was assigned four topics, two for searching and two for browsing. One topic of each task was carried out using each of the presentation modes. The order of the topics and presentation modes were randomized using a Graeco-Latin square design such that 1) half of the participants started with the list mode and another half with the visual mode; and, 2) half of the

participants started with a searching topic and another half with a browsing topic. According to our pre-tests of the procedure, the participants were given 9 minutes for each task. A logging software tool, QT Honey 1 was employed to record all user interactions with the system including clicks, scrolls, mouse focuses, etc.

users to find “as many songs as possible”. Although the difference was a little shy from being significant, users did view more pages using the visual mode than using the list mode, which suggests greater user efforts paid on the visual mode, or, as discussed below, users might just have liked to play with the visual mode. Table 3. Means and standard derivations (in parentheses) of user effectiveness and user perception measures. Measure List Visual p value number of songs found 6.11 (2.50) 5.69 (2.47) 0.05* task completion time 8.64 (0.58) 8.57 (0.74) 0.22 number of songs played 22.16 (8.37) 21.2 (7.87) 0.28 number of pages viewed 4.42 (3.70) 4.90 (3.74) 0.13 task easiness 4.55 (1.63) 4.03 (1.80) <0.01* satisfaction 4.59 (1.62) 4.25 (1.64) 0.06* N = 32, p < 0.1, the unit of task completion time is minute.

Figure 3. Distribution of frequencies the participants listen to and search for music.

As there were two tasks, it is useful to find out whether the two presentation modes have relative advantages in each of the tasks. Table 4 presents means and standard derivations of the above measures regarding to presentation modes and task types.

After finishing all the four topics, a group interview was conducted in Japanese with the participants in each batch. It focused on users’ detailed opinions regarding to the two presentation modes of moodydb. The entire procedure lasted about 1.5 hours and each participant was paid 2000 yen for their participation.

Table 4. Measures w.r.t. presentation modes and task types.

3.5 Measures The following user-centered measures were used to compare the two presentation modes of moodydb.

Mean (std) Searching Browsing

User effectiveness: 1) number of songs found; 2) task completion time; 3) number of songs played; and, 4) number of result pages viewed. The first measure was collected from the answer sheets where participants wrote down the songs they found. The other three measures were collected from the system logs.

Mean (std) Searching Browsing

User perception: 1) task easiness; 2) satisfaction with the songs found. These two measures were collected from the answer sheets. Both were on a Likert scale from 1 to 7. (1: very difficult/very dissatisfied; 7: very easy/very satisfied).

Mean (std) Searching Browsing

User preference: which presentation mode a user prefers. The measures were collected from the following questions in the postexperiment questionnaire: a) In general, which presentation mode do you like more? b) Which presentation makes the searching tasks easier? c) Which presentation makes the browsing tasks easier? d) Which presentation mode is more visually appealing? e) Which presentation mode is more fun?

Mean (std) Searching Browsing

4. PRELIMINARY RESULTS Table 3 presents user effectiveness and user perception measures of the two presentation modes. There were significant differences between the two modes on three of the measures at p < 0.1 level. Specifically, Using the list mode, users found more songs, felt the tasks were easier, and were more satisfied with the songs they found than using the visual mode. In terms of task completion time, users generally used up the given 9 minutes or worked until very close to the end. This probably is because the tasks asked 1

http://cres.jpn.org/?QT-Honey

Mean (std) Searching Browsing

Mean (std) Searching Browsing N = 32, p < 0.1

number of songs found List Visual 5.47 (2.30) 4.88 (2.21) 6.75 (2.55) 6.5 (2.49) task completion time (minutes) List Visual 8.74 (0.36) 8.62 (0.37) 8.60 (0.67) 8.48 (0.61) number of songs played List Visual 23.65 (9.28) 23.50 (8.16) 20.68 (7.19) 19.00 (7.01) number of page viewed List Visual 4.32 (3.59) 5.10 (3.92) 4.52 (3.85) 4.71 (3.63) task easiness List Visual 3.94 (1.64) 3.41 (1.83) 5.16 (1.39) 4.66 (1.56) satisfaction of the answers List Visual 4.00 (1.63) 3.34 (1.47) 5.19 (1.40) 5.16 (1.27)

p value 0.05* 0.27 p value 0.12 0.29 p value 0.38 0.09* p value 0.08* 0.39 p value 0.01* 0.06* p value 0.03* 0.46

Table 4 shows the two modes made more differences on the searching task than on the browsing task. In particular, for the searching task, users found more answers and were more satisfied using the list mode while such differences on the browsing task were not significant. Users perceived both tasks were easier when using the list mode, but the difference was more significant for the searching task than the browsing one. All these evidences support that the list mode had advantages in searching tasks. As the system was not as familiar to the users as general search engines, we examined the impact of topic order on the measure,

number of songs found. As shown in Table 5, the two presentation modes had not much difference on the first searching topic, while for the second searching topic users found significantly more songs using the list mode. This indicates the list mode were more effective for the searching task but required users’ learning efforts. For the browsing task, the visual mode was more effective on the first topic while the list mode was more effective on the second. This might be related to the specific mood categories selected by the participants when they performed the browsing topics, but further analysis is needed to better understand the observation. Table 5. number of songs found w.r.t task order. Mean (std) 1st searching 2nd searching 1st browsing 2nd browsing N = 32, p < 0.1

List 4.69 (2.02) 6.25 (2.35) 5.19(1.87) 8.31 (2.18)

Visual 5.31 (2.57) 4.44 (1.75) 7.25 (2.41) 5.75 (2.41)

p value 0.23 0.01* <0.01* <0.01*

Figure 3 shows the distribution of users’ preferences between the two presentation modes. It shows that more users preferred the list mode for both tasks while in general each mode was preferred by nearly equal number of users. This indicates that users had other criteria than task supporting alone in evaluating MIR interfaces. This is supported by the interview data: some participants explained that they preferred the visual mode because “it was fun”.

when an interface is visually appealing and fun, it is possible that users become less concerned about their initial information needs and instead start enjoying the interface itself. This is especially applicable to music information seeking where the main motivation is entertainment rather than task-driven. An important implication for interactive IR evaluation is that measures focusing on task completion may not be sufficient and can be supplemented by measures on engagement and enjoyment.

5. CONCLUSIONS AND FUTURE WORK This paper presents a user experiment evaluating two result presentation modes of a music mood recommendation system on two task types, searching and browsing. Preliminary results show significant differences between the two modes on user-centered measures. More data analysis is needed to further our understanding on the relative advantages of the two presentation modes on each task as well as users’ interactive behaviors in seeking for music information and underlying reasons.

6. ACKNOWLEDGMENTS The research is partially supported by the JSPS Grand-in-Aid (#21300096) and the Non-MOU Grant funded by the National Institute of Informatics in Japan. We thank our anonymous reviewers for their invaluable suggestions.

7. REFERENCES [1] Downie, J. S. 2008. The Music Information Retrieval Evaluation eXchange (2005–2007): A window into music information retrieval research. Acoustical Science and Technology, 29(4), 247–255.

[2] Hoashi, K., Hamawaki, S., Ishizaki, H., Takishima, Y. and Katto, J. 2009. Usability Evaluation of Visualization Interfaces for ContentBased Music Retrieval Systems, In Proceedings of the 5th International Conference on Music Information Retrieval (ISMIR).

[3] Hu, X, Downie, J. S., Laurier, C., Bay, M. and Ehmann, A. F. 2008. The 2007 MIREX Audio Mood Classification Task: Lessons Learned. ISMIR. pp. 462-467

[4] Hu, X., Sanghvi, V., Vong, B., On, P. J., Leong, C., and Angelica, J. 2008. MOODY: A Web-Based Music Mood Classification and Recommendation System, ISMIR.

[5] Juslin, P. N. and Sloboda, J. A. 2001. Music and emotion: introduction. In P. N. Juslin and J. A. Sloboda (Eds.), Music and Emotion: Theory and Research. New York: Oxford University Press.

Figure 4. User preferences between the two presentation modes. The group interviews also revealed: 1) participants liked the list mode because the ranked order made it easier for them to keep track of songs they had examined, which was helpful for recall oriented tasks (“to find as many songs as possible”); 2) participants liked the visual mode because it was fun and were good for exploratory tasks (“to browse”, “to look around”); 3) many participants expressed positive opinions on the graphic representation of songs using the album cover on the visual mode, especially for browsing task. They related this to their experience of “Jacket Gai” (A Japanese term meaning to choose albums by the impression of their cover only). The observation that users were less effective using the visual mode but liked it nonetheless seems to reflect the “user-distraction” phenomenon where users express satisfaction with irrelevant search results [9]. The interview results further corroborate that

[6] Lamere, P. and Eck, D. 2007. Using 3D visualizations to explore and discover music, ISMIR.

[7] Lee, J. H. and Downie, J. S. 2004. Survey of music information needs, uses, and seeking behaviours: preliminary findings. ISMIR, pp. 441-446.

[8] Pampalk, E., Rauber, A. and Merkl, D. 2002. Content-based Organization and Visualization of Music Archives, In the Proceedings of the 10th ACM International Conference on Multimedia (MM'02), pp 570-579.

[9] Soergel, D. 1976. Is user satisfaction a hobgoblin? Journal of the American Society for Information Science, 27 (4), 256-259.

[10] Vignoli, F. 2004. Digital music interaction concepts: a user study. ISMIR.

[11] Witten, I. and Frank, E. 1999. Data mining: Practical machine learning tools and techniques with Java implementations. San Fransisco: Morgan Kaufmann.

User Evaluation of an Interactive Music Information Retrieval System

H.3.3 [Information Storage and Retrieval]: Information Search and Retrieval – search ... To date, most MIR systems and online services have been using the ...

653KB Sizes 0 Downloads 230 Views

Recommend Documents

An Architectural Framework for Interactive Music Systems
Software Architecture, Interactive Systems, Music soft- ... synthesis of data media of different nature. ... forms (e.g. Max/MSP [19] and Pure Data [24]), and oth-.

EFFICIENT INTERACTIVE RETRIEVAL OF SPOKEN ...
between the key term ti and the corresponding document class C(ti) is defined by .... initially large number of users can be further classified into cate- gories by ...

Image retrieval system and image retrieval method
Dec 15, 2005 - face unit to the retrieval processing unit, image data stored in the image information storing unit is retrieved in the retrieval processing unit, and ...

International Society for Music Information Retrieval -
Call for Financial Support. The ​ISMIR Society​, ​Women in MIR​, and local organizers are offering a number of financial support opportunities for members of ...

Improving Arabic Information Retrieval System using n-gram method ...
Improving Arabic Information Retrieval System using n-gram method.pdf. Improving Arabic Information Retrieval System using n-gram method.pdf. Open. Extract.

Method of wireless retrieval of information
Dec 24, 1996 - mitted to the database service provider which then provides a response back to .... a CNI, an e-mail address, a fax number, etc. and any of this.

Method of wireless retrieval of information
Dec 24, 1996 - provider can establish a database which can be accessed by subscribers to ... message center then forwards the messages to the designated mobile station via ... tion to place a call via the wireless network and a PSTN to an.

information storage and retrieval system pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. information ...

A Content-Based Information Retrieval System for ... - Springer Link
This paper deals with the elaboration of an interactive software which ... Springer Science + Business Media B.V. 2008 .... Trend boards offer a good representation of the references used ..... function could be fulfilled by mobile devices.

Evaluation of an Ontology-based Knowledge-Management-System. A ...
Fur- thermore, the system enables the efficient administration of large amounts of data in accordance with a knowledge management system and possesses the ...

Evaluation of an automated furrow irrigation system ...
crop (63.14 kg/ha/cm) was considerably higher than of conventional method (51.43 kg/ha/cm). Key words ... no need to go to the field at night or any other ...

an effective video retrieval system by combining visual ...
May 2, 2009 - COMBINING VISUAL AND TEXTUAL MINING .... Totally, the duration of video data is about 20 hours ... Recall = 5 / 20 * 100% = 25%. %100*.

An Evaluation of a Collision Handling System using ...
A number of experiments in virtual scenarios with objects falling in a static plane ... Realism—Animation; I.3.7 [Virtual Reality]: Three-Di- mensional Graphics ..... Physical Modeling, pages 173 – 184, Cardiff, Wales,. UK, 2006. ACM Press. 1245.

An Evaluation of Layout Features for Information ...
is O (meaning “not belonging to any entity”) and the .... dictionary features and are case-insensitive) ... remove duplicate and near duplicate messages (based on.

SemRetriev – an Ontology Driven Image Retrieval System
Jul 9, 2007 - prototype system for image retrieval which combines the use of an ontology which structures an ... SemRetriev is meant to show that the use of semantic .... After the elimination of invalid links and invalid files, there are around ...

Information Retrieval with Actions and Change: an ASP ...
In this paper, we advance this line of research by considering the case in ... develop needed mathematical foundations and propose a formalization of the ... ios. Finally, we briefly present related work, draw conclusions and discuss future work.

An Interactive Augmented Reality System for Engineering Education
virtual museum, which through the use of technologies such as Web/X3D, VR and AR offers to the ... for new virtual tours simply by replacing repository content.

Implementing an Interactive Visualization System on a ...
Currently, programming such systems requires that algorithms be ...... [CAM94] W. J. Camp, S. J. Plimpton, B. A. Hendrickson, and R. W. Leland, “Massively ...

Implementing an Interactive Visualization System on a ... - CiteSeerX
Department of Computer Science. University of Massachusetts-Lowell. One University Avenue. Lowell, MA 01854. Abstract. The use of a massively parallel ...

Implementing an Interactive Visualization System on a ... - CiteSeerX
formed by connecting line segments at their endpoints according to a set of geometric ... supercomputer version of the color icon and briefly outline the benefits and ..... The memory limitations of the Terasys forces the use of memory to be an impor

An Interactive Augmented Reality System for Engineering Education
The implemented framework is composed an XML data repository, an XML ... virtual and AR in the same web based learning support .... The software architecture of ARIFLite is ... environment or a seminar room, the student would launch.