ISCA - International Speech
Communication Association


ISCApad Archive  »  2013  »  ISCApad #180  »  Events  »  Other Events  »  Interspeech 2013 ComParE: Computational Paralinguistics Challenge

ISCApad #180

Monday, June 10, 2013 by Chris Wellekens

3-3-13 Interspeech 2013 ComParE: Computational Paralinguistics Challenge
  

Call for Participation

INTERSPEECH 2013 ComParE:COMPUTATIONAL PARALINGUISTICS CHALLENGE

  Social Signals, Conflict, Emotion, Autism

Fourth Sub-Challenge now open – obtain the data:

http://emotion-research.net/sigs/speech-sig/IS13-Challenge-Agreements-SC2.pdf

 

The Challenge

After four consecutive Challenges at INTERSPEECH, there still exists a multiplicity of not yet covered, but highly relevant paralinguistic phenomena. In the last instalments, we focused on single speakers. With a new task, we now want to broaden to analysing discussion of multiple speakers in the Conflict Sub-Challenge. A further novelty is introduced by the Social Signals Sub-Challenge: For the first time, non-linguistic events have to be classified and localised – laughter and fillers. In the Emotion Sub-Challenge we are literally “going back to the roots”. However, by intention, we use acted material for the first time to fuel the ever on-going discussion on differences between naturalistic and acted material and hope to highlight the differences. Finally, the Autism Sub-Challenge picks up on Autism Spectrum Condition in children’s speech in this year. Apart from intelligent and socially competent future agents and robots, main applications are found in the medical domain and surveillance. The Challenge corpora feature rich annotation such as speaker meta-data, orthographic transcript, phonemic transcript, and segmentation. All four are given with distinct definitions of test, development, and training partitions, incorporating speaker independence as needed in most real-life settings. Benchmark results of the most popular approaches will be provided as in the years before. In these respects, the INTERSPEECH 2013 COMPUTATIONAL PARALINGUISTICS CHALLENGE (ComParE) shall help bridging the gap between excellent research on paralinguistic information in spoken language and low compatibility of results.

 

In summary, four Sub-Challenges are addressed:

 

•             In the Social Signals Sub-Challenge, non-linguistic events – laughter and fillers – of a speaker have to be classified and localised based on acoustics.

•             In the Conflict Sub-Challenge, group discussions have to be automatically evaluated aiming at retrieving conflicts.

•             In the Emotion Sub-Challenge, the emotion of a speaker’s voice has to be determined by a suited learning algorithm and acoustic features.

•             In the Autism Sub-Challenge, the type of pathology of a speaker has to be determined by a suited classification algorithm and acoustic features.

 

The measures of competition will be Unweighted Average Area Under receiver operating Curve and Recall. All Sub-Challenges allow contributors to find their own features with their own machine learning algorithm. However, a standard feature set will be provided per corpus that may be used. Participants will have to stick to the definition of training, development, and test sets. They may report on results obtained on the development set, but have only five trials to upload their results on the test sets, whose labels are unknown to them. Each participation will be accompanied by a paper presenting the results that undergoes peer-review and has to be accepted for the conference in order to participate in the Challenge. The organisers preserve the right to re-evaluate the findings, but will not participate themselves in the Challenge. Participants are encouraged to compete in all Sub-Challenges.

 

Overall, contributions using the provided or equivalent data are sought for (but not limited to):

 

•             Participation in a Sub-Challenge

•             Contributions focussing on Computational Paralinguistics centred around the Challenge topics

 

The results of the Challenge will be presented at Interspeech 2013 in Lyon, France. Prizes will be awarded to the Sub-Challenge winners. If you are interested and planning to participate in INTERSPEECH 2013 ComParE, or if you want to be kept informed about the Challenge, please send the organisers an e-mail to indicate your interest and visit the homepage: http://emotion-research.net/sigs/speech-sig/is13-compare

 

 

Organisers:

 

Björn Schuller (TUM, Germany)

Stefan Steidl (FAU Erlangen-Nuremberg, Germany)

Anton Batliner (TUM, Germany)

Alessandro Vinciarelli (University of Glasgow, UK)

Klaus Scherer (Swiss Center for Affective Sciences, Switzerland)

Fabien Ringeval (University of Fribourg, Switzerland)

Mohamed Chetouani (Université Pierre et Marie Curie, France)

 

 

Dates:

 

Paper Submission            18 March             2013

Final Result Upload         24 May                 2013

Camera-ready Paper      29 May                 2013

 

 

Sponsors:

 

HUMAINE Association                   (http://emotion-research.net/)

SSPNet                                                 (http://sspnet.eu/)

ASC-Inclusion                                    (http://www.asc-inclusion.eu/)

 

___________________________________________

 

PD Dr. habil. DI Björn W. Schuller

 

Head Machine Intelligence & Signal Processing Group

Institute for Human-Machine Communication

Technische Universität München

D-80333 München

Germany

 

+49-(0)89-289-28548

 

schuller@tum.de

www.mmk.ei.tum.de/~sch


Back  Top


 Organisation  Events   Membership   Help 
 > Board  > Interspeech  > Join - renew  > Sitemap
 > Legal documents  > Workshops  > Membership directory  > Contact
 > Logos      > FAQ
       > Privacy policy

© Copyright 2024 - ISCA International Speech Communication Association - All right reserved.

Powered by ISCA