MediaEval 2015

MediaEval 2015 Workshop
The MediaEval 2015 Workshop took place in Wurzen, Germany, on 14-15 September 2015 as a satellite event of Interspeech 2015. The workshop brought together task participants to present and discuss their findings, and prepare for future work. A list of the tasks offered in MediaEval 2015 is below. Workshop materials and media are available online:

How to cite the proceedings
Martha Larson, Bogdan Ionescu, Mats Sjöberg, Xavier Anguera, Johann Poignant, Michael Riegler, Maria Eskevich, Claudia Hauff, Richard Sutcliffe, Gareth J.F. Jones, Yi-Hsuan Yang, Mohammad Soleymani, Symeon Papadopoulos (eds.), Proceedings of the MediaEval 2015 Multimedia Benchmark Workshop, Wurzen, Germany, September 14-15, 2015,, online

How to cite a paper (example)
Petra Galuščáková, Pavel Pecina, CUNI at MediaEval 2015 Search and Anchoring in Video Archives: Anchoring via Information Retrieval, Proceedings of the MediaEval 2015 Workshop, Wurzen, Germany, September 14-15, 2015,, online

MediaEval 2015 Organizers
  • For the complete list of task organizers the formed the over all organization of the MediaEval 2015 benchmark, please see "Who are we?" and also the individual task pages.
  • For a list of the people whose organizational effort made the workshop possible, please see below.

MediaEval 2015 Tasks
The following tasks were offered in MediaEval 2015.

QUESST: Query by Example Search on Speech Task
The task involves searching FOR audio content WITHIN audio content USING an audio content query. This task is particularly interesting for speech researchers in the area of spoken term detection or low-resource/zero-resource speech processing. The primary performance metric will be the normalized cross entropy cost (Cnxe). Read more...

Multimodal Person Discovery in Broadcast TV (New in 2015!)
Given raw TV broadcasts, each shot must be automatically tagged with the name(s) of people who can be both seen as well as heard in the shot. The list of people is not known a priori and their names must be discovered in an unsupervised way from provided text overlay or speech transcripts. The task will be evaluated on a new French corpus (provided by INA) and the AGORA Catalan corpus, using standard information retrieval metrics based on a posteriori collaborative annotation of the corpus. Read more...

C@merata: Querying Musical Scores
The input is a natural language phrase referring to a musical feature (e.g., ‘consecutive fifths’) together with a classical music score, and the required output is a list of passages in the score which contain that feature. Scores are in the MusicXML format, which can capture most aspects of Western music notation. Evaluation is via versions of Precision and Recall relative to a Gold Standard produced by the organisers. Read more...

Affective Impact of Movies (including Violent Scenes Detection)
In this task participating teams are expected to classify short movie scenes by their affective content according to two use cases: (1) the presence of depicted violence, and (2) their emotional impact (valence, arousal). The training data consists of short Creative Commons-licensed movie scenes (both professional and amateur) together with human annotations of violence and valence-arousal ratings. The results will be evaluated using standard retrieval and classification metrics. Read more...

Emotion in Music (An Affect Task)
We aim at detecting emotional dynamics of music using its content. Given a set of songs, participants are asked to automatically generate continuous emotional representations in arousal and valence. Read more...

Retrieving Diverse Social Images
This task requires participants to refine a ranked list of Flickr photos with location related information using provided visual, textual and user credibility information. Results are evaluated with respect to their relevance to the query and the diverse representation of it. Read more...

Placing: Multimodal Geo-location Prediction
The Placing Task requires participants to estimate the locations where multimedia items (photos or videos) were captured solely by inspecting the content and metadata of these items, and optionally exploiting additional knowledge sources such as gazetteers. Performance is evaluated using the distance to the ground truth coordinates of the multimedia items. Read more...

Verifying Multimedia Use (New in 2015!)
For this task, the input is a tweet about an event that has the profile to be of interest in the international news, and the accompanying multimedia item (image or video). Participants must build systems that output a binary decision representing a verification of whether the multimedia item reflects the reality of the event in the way purported by the tweet. The task is evaluated using the F1 score. Participants are also requested to return a short explanation or evidence for the verification decision. Read more...

Context of Experience: Recommending Videos Suiting a Watching Situation (New in 2015!)
This task develops multimodal techniques for automatic prediction of multimedia in a particular consumption content. In particular, we focus on the context of predicting movies that are suitable to watch on airplanes. Input to the prediction methods is movie trailers, and metadata from IMDb. Output is evaluated using the Weighted F1 score, with expert labels as ground truth. Read more...

Synchronization of Multi-User Event Media
This task addresses the challenge of automatically creating a chronologically-ordered outline of multiple multimedia collections corresponding to the same event. Given N media collections (galleries) taken by different users/devices at the same event, the goal is to find the best (relative) time alignment among them and detect the significant sub-events over the whole gallery. Performance is evaluated using ground truth time codes and actual event schedules. Read more...

DroneProtect: Mini-drone Video Privacy Task (New in 2015!)
Recent popularity of mini-drones and their rapidly increasing adoption in various areas, including photography, news reporting, cinema, mail delivery, cartography, agriculture, and military, raises concerns for privacy protection and personal safety. Input to the task is drone video, and output is version of the video which protects privacy while retaining key information about the event or situation recorded. Read more...

Search and Anchoring in Video Archives
The 2015 Search and Anchoring in Video Archives task consists of two sub-tasks: search for multimedia content and automatic anchor selection. In the “search for multimedia content” sub-task, participants use multimodal textual and visual descriptions of content of interest to retrieve potentially relevant video segments from within a collection. In the “automatic anchor selection” sub-task, participants automatically predict key elements of videos as anchor points for the formation of hyperlinks to relevant content within the collection. The video collection consists of professional broadcasts from BBC or semi-professional user generated content. Participant submissions will be assessed using professionally-created anchors, and crowdsourcing-based evaluation. Read more...

MediaEval 2015 Community Council
Mohammad Soleymani (University of Geneva, Switzerland)
Bogdan Ionescu (Politehnica of Bucharest, Romania)
Guillaume Gravier (IRISA, France)
Gareth Jones (Dublin City University, Ireland)
Martha Larson (Delft University of Technology, Netherlands) (Contact and Workshop General Chair)


Key contributors to 2015 organization
Saskia Peters (Delft University of Technology, Netherlands)
Michael Riegler, (Simula Research Lab AS, Norway)
Bogdan Boteanu (University Politehnica of Bucharest, Romania)
Richard Sutcliffe (University of Essex, UK)

Sponsors and Supporters:
MediaEval 2015 thanks these organizations for their sponsorship and support:

Pasted Graphiclogo
page37-logo-4Elias Logo



Technical Committee TC12 "Multimedia and Visual Information Systems"
of the International Association of Pattern Recognition

For information on how to become a sponsor or supporter of MediaEval 2015, please contact Martha Larson m (dot) a (dot) larson (at)