| --- INTERSPEECH 2014 - SINGAPORE
--- September 14-18, 2014
--- http://www.INTERSPEECH2014.org
The INTERSPEECH 2014 Organising Committee is pleased to announce
the following 8 tutorials presented by distinguished speakers
at the conference and will be offered on Sunday, 14 September 2014.
All Tutorials will be of three (3) hours duration and require
an additional registration fee (separate from the conference registration fee).
• Non-speech acoustic event detection and classification
• Contribution of MRI to Exploring and Modeling Speech Production
• Computational Models for Audiovisual Emotion Perception
• The Art and Science of Speech Feature Engineering
• Recent Advances in Speaker Diarization
• Multimodal Speech Recognition with the AusTalk 3D Audio-Visual Corpus
• Semantic Web and Linked Big Data Resources for Spoken Language Processing
• Speech and Audio for Multimedia Semantics
----------------------------------------------------------------------------------------------------
ISCSLP Tutorials @ INTERSPEECH 2014
----------------------------------------------------------------------------------------------------
Additionally, the ISCSLP 2014 Organising Committee welcomes
the INTERSPEECH 2014 delegates to join the 4 ISCSLP tutorials
which will be offered on Saturday, 13 September 2014.
• Adaptation Techniques for Statistical Speech Recognition
• Emotion and Mental State Recognition: Features, Models, System Applications and Beyond
• Unsupervised Speech and Language Processing via Topic Models
• Deep Learning for Speech Generation and Synthesis
More information available at: http://www.interspeech2014.org/public.php?page=tutorial.html
----------------------------------------------------------------------------------------------------
Tutorials Description
----------------------------------------------------------------------------------------------------
T1: Non-speech acoustic event detection and classification
The research in audio signal processing has been dominated by speech research,
but most of the sounds in our real-life environments are actually non-speech
events such as cars passing by, wind, warning beeps, and animal sounds.
These acoustic events contain much information about the environment and physical
events that take place in it, enabling novel application areas such as safety,
health monitoring and investigation of biodiversity. But while recent years
have seen wide-spread adoption of applications such as speech recognition and
song recognition, generic computer audition is still in its infancy.
Non-speech acoustic events have several fundamental differences to speech,
but many of the core algorithms used by speech researchers can be leveraged
for generic audio analysis. The tutorial is a comprehensive review of the field
of acoustic event detection as it currently stands. The goal of the tutorial is
foster interest in the community, highlight the challenges and opportunities
and provide a starting point for new researchers. We will discuss what acoustic
event detection entails, the commonalities differences with speech processing,
such as the large variation in sounds and the possible overlap with other sounds.
We will then discuss basic experimental and algorithm design, including descriptions
of available databases and machine learning methods. We will then discuss more
advanced topics such as methods to deal with temporally overlapping sounds and
modelling the relations between sounds. We will finish with a discussion of
avenues for future research.
Organizers: Tuomas Virtanen and Jort F. Gemmeke
T2: Contribution of MRI to Exploring and Modeling Speech Production
Magnetic resonance imaging (MRI) provides us a magic vision to look into
the human body in various ways not only with static imaging but also with
motion imaging. MRI has been a powerful technique for speech research to
study finer anatomy of the speech organs or to visualize true vocal tracts
in three dimensions. Inherent problems of slow image acquisition for speech
tasks or insufficient signal-to-noise ratio for microscopic observation have
been the cost for researchers to search for task-specific imaging techniques.
The recent advances of the 3-Tesla technology suggest more practical solutions
to broader applications of MRI by overcoming previous technical limitations.
In this joint tutorial in two parts, we summarize our previous effort to accumulate
scientific knowledge with MRI and to advance speech modeling studies for future
development. Part I, given by Kiyoshi Honda, introduces how to visualize the
speech organs and vocal tracts by presenting techniques and data for finer static
imaging, synchronized motion imaging, surface marker tracking, real-time imaging,
and vocal-tract mechanical modeling. Part 2, presented by Jianwu Dang, focuses on
applications of MRI for phonetics of Mandarin vowels, acoustics of the vocal tracts
with side branches, analysis and simulation in search of talker characteristics,
physiological modeling of the articulatory system, and motor control paradigm
for speech articulation.
Organizers: Kiyoshi HONDA and Jianwu DANG
T3: Computational Models for Audiovisual Emotion Perception
In this tutorial we will explore engineering approaches to understanding human
emotion perception, focusing both on modeling and application. We will highlight
both current and historical trends in emotion perception modeling, focusing on
both psychological and engineering-driven theories of perception
(statistical analyses, data-driven computational modeling, and implicit sensing).
The importance of this topic can be appreciated from both an engineering viewpoint,
any system that either models human behavior or interacts with human partners must
understand emotion perception as it fundamentally underlies and modulates our
communication, or from a psychological perspective, emotion perception is also used
in the diagnosis of many mental health conditions and is tracked in therapeutic
interventions. Research in emotion perception seeks to identify models that describe
the felt sense of ‘typical’ emotion expression – i.e., an observer/evaluator’s attribution
of the emotional state of the speaker. This felt sense is a function of the methods through
which individuals integrate the presented multimodal emotional information.
We will cover psychological theories of emotion, engineering models of emotion,
and experimental approaches to measure emotion. We will demonstrate how these modeling
strategies can be used as a component of emotion classification frameworks and how
they can be used to inform the design of emotional behaviors.
Organizers: Emily Mower Provost and Carlos Busso
T4: The Art and Science of Speech Feature Engineering
With significant advances in mobile technology and audio sensing devices,
there is a fundamental need to describe vast amounts of audio data in terms
of well representative lower dimensional descriptors for efficient automatic
processing. The extraction of these signal representations, also called features,
constitutes the first step in processing a speech signal. The art and science of
feature engineering relates to addressing the two inherent challenges - extracting
sufficient information from the speech signal for the task at hand and suppressing
the unwanted redundancies for computational efficiency and robustness.
The area of speech feature extraction combines a wide variety of disciplines like
signal processing, machine learning, psychophysics, information theory, linguistics and physiology.
It has a rich history spanning more than five decades and has seen tremendous advances
in the last few years. This has propelled the transition of the speech technology from
controlled environments to millions of end user applications.
In this tutorial, we review the evolution of speech feature processing methods,
summarize the recent advances of the last two decades and provide insights into the
future of feature engineering. This will include the discussions on the spectral
representation methods developed in the past, human auditory motivated techniques
for robust speech processing, data driven unsupervised features like ivectors and
recent advances in deep neural network based techniques. With experimental results,
we will also illustrate the impact of these features for various state-of-the-art
speech processing systems. The future of speech signal processing will need to address
various robustness issues in complex acoustic environments while being able
to derive useful information from big data.
Organizers: Sriram Ganapathy and Samuel Thomas
T5: Recent Advances in Speaker Diarization
The tutorial will start with an introduction to speaker diarization giving a general
overview of the subject. Afterwards, we will cover the basic background including
feature extraction, and common modeling techniques such as GMMs and HMMs.
Then, we will discuss the first processing step usually done in speaker diarization
which is voice activity detection. We will consequently describe the classic approaches
for speaker diarization which are widely used today. We will then introduce state-of-the-art
techniques in speaker recognition required to understand modern speaker diarization techniques.
Following, we will describe approaches for speaker diarization using advanced representation
methods (supervectors, speaker factors, i-vectors) and we will describe supervised and
unsupervised learning techniques used for speaker diarization. We will also discuss issues
such as coping with unknown number of speakers, detecting and dealing with overlapping speech,
diarization confidence estimation, and online speaker diarization. Finally we will discuss
two recent works: exploiting a-prioiri acoustic information (such as processing a meeting
when some of the participants are known in advanced to the system, and training data is available for them),
The second recent work is modeling speaker-turn dynamics. If time permits, we will also discuss concepts
such as multi-modal diarization and using TDOA (time difference of arrival) for diarization of meetings.
Organizers: Hagai Aronowitz
T6: Multimodal Speech Recognition with the AusTalk 3D Audio-Visual Corpus
This tutorial will provide attendees a brief overview of 3D based AVSR research.
In this tutorial, attendees will learn how to use the newly developed 3D based audio
visual data corpus we derived from the AusTalk corpus (https://austalk.edu.au/)
for audio-visual speech/speaker recognition. In addition, we also plan to introduce
some results using this newly developed 3D audio-visual data corpus, which show that
there is a significant speech accuracy increase by integrating both depth-level and grey-level
visual features. In the first part of the tutorial, we will review some recent works published
in the last decade, so that attendees can obtain an overview of the fundamental concepts
and challenges in this field. In the second part of the tutorial, we will briefly describe
the recording protocol and contents of the 3D data corpus, and show attendees how to use
this corpus for their own research. In the third part of this tutorial, we will present our
results using the 3D data corpus. The experimental results show that, compared with the
conventional AVSR based on the audio and grey-level visual features, the integration of grey
and depth visual information can boost the AVSR accuracy significantly. Moreover,
we will also experimentally explain why adding depth information can benefit the standard AVSR systems.
Eventually, through our tutorial, we hope we can inspire more researchers in the community
to contribute to this exciting research.
Organizers: Roberto Togneri, Mohammed Bennamoun and Chao (Luke) Sui
T7: Semantic Web and Linked Big Data Resources for Spoken Language Processing
State-of-the-art statistical spoken language processing typically requires
significant manual effort to construct domain-specific schemas (ontologies)
as well as manual effort to annotate training data against these schemas.
At the same time, a recent surge of activity and progress on semantic web-related
concepts from the large search-engine companies represents a potential alternative
to the manually intensive design of spoken language processing systems.
Standards such as schema.org have been established for schemas (ontologies) that
webmasters can use to semantically and uniformly markup their web pages.
Search engines like Bing, Google, and Yandex have adopted these standards and are
leveraging them to create semantic search engines at the scale of the web.
As a result, the open linked data resources and semantic graphs covering various
domains (such as Freebase [3]) have grown massively every year and contains far more
information than any single resource anywhere on the Web. Furthermore, these resources
contain links to text data (such as Wikipedia pages) related to the knowledge in the graph.
Recently, several studies on speech language processing started exploiting these massive
linked data resources for language modeling and spoken language understanding.
This tutorial will include a brief introduction to the semantic web and the linked
data structure, available resources, and querying languages.
An overview of related work on information extraction and language processing will
be presented, where the main focus will be on methods for learning spoken language
understanding models from these resources.
Organizers: Dilek Hakkani-Tür and Larry Heck
T8: Speech and Audio for Multimedia Semantics
Internet media sharing sites and the one-click upload capability of smartphones
are producing a deluge of multimedia content. While visual features are often dominant
in such material, acoustic and speech information in particular often complements it.
By facilitating access to large amounts of data, the text-based Internet gave a huge
boost to the field of natural language processing. The vast amount of consumer-produced
video becoming available now will do the same for video processing, eventually enabling
semantic understanding of multimedia material, with implications for human computer interaction, robotics, etc.
Large-scale multi-modal analysis of audio-visual material is now central to a number of
multi-site research projects around the world. While each of these have slightly different
targets, they are facing largely the same challenges: how to robustly and efficiently process
large amounts of data, how to represent and then fuse information across modalities,
how to train classifiers and segmenters on unlabeled data, how to include human feedback, etc.
In this tutorial, we will present the state of the art in large-scale video, speech,
and non-speech audio processing, and show how these approaches are being applied to tasks
such as content based video retrieval (CBVR) and multimedia event detection (MED).
We will introduce the most important tools and techniques, and show how the combination of
information across modalities can be used to induce semantics on multimedia material
through ranking of information and fusion. Finally, we will discuss opportunities
for research that the INTERSPEECH community specifically will find interesting and fertile.
Organizers: Florian Metze and Koichi Shinoda
----------------------------------------------------------------------------------------------------
ISCSLP Tutorials @ INTERSPEECH 2014 Description
----------------------------------------------------------------------------------------------------
ISCSLP-T1: Adaptation Techniques for Statistical Speech Recognition
Adaptation is a technique to make better use of existing models for test data
from new acoustic or linguistic conditions. It is an important and challenging
research area of statistical speech recognition. This tutorial gives a systematic
review of fundamental theories as well as introduction of state-of-the-art adaptation
techniques. It includes both acoustic and language model adaptation. Following a simple example
of acoustic model adaptation, basic concepts, procedures and categories of adaptation will
be introduced. Then, a number of advanced adaptation techniques will be discussed,
such as discriminative adaptation, Deep Neural Network adaptation, adaptive training,
relationship to noise robustness etc. After the detailed review of acoustic model adaptation,
an introduction of language model adaptation, such as topic adaptation will also be given.
The whole tutorial is then summarised and future research direction will be discussed.
Organizers: Kai Yu
ISCSLP-T2: Emotion and Mental State Recognition: Features, Models, System Applications and Beyond
Emotion recognition is the ability to identify what you are feeling from moment
to moment and to understand the connection between your feelings and your expressions.
In today’s world, human-computer interaction (HCI) interface undoubtedly plays an
important role in our daily life. Toward harmonious HCI interfaces, automated analysis
and recognition of human emotion has attracted increasing attention from researchers
in multidisciplinary research fields. A specific area of current interest that also has key
implications for HCI is the estimation of cognitive load (mental workload), research into
which is still at an early stage. Technologies for processing daily activities including speech,
text and music have expanded the interaction modalities between humans and computer-supported
communicational artifacts.
In this tutorial, we will present theoretical and practical work offering new and broad views
of the latest research in emotional awareness from audio and speech. We discuss several parts
spanning a variety of theoretical background and applications ranging from salient emotional features,
emotional-cognitive models, compensation methods for variability due to speaker and linguistic content,
to machine learning approaches applicable to emotion recognition. In each topic, we will review
the state of the art by introducing current methods and presenting several applications.
In particular, the application to cognitive load estimation will be discussed,
from its psychophysiological origins to system design considerations. Eventually,
technologies developed in different areas will be combined for future applications,
so in addition to a survey of future research challenges,
we will envision a few scenarios in which affective computing can make a difference.
Organizers: Chung-Hsien Wu, Hsin-Min Wang, Julien Epps and Vidhyasaharan Sethu
ISCSLP-T3: Unsupervised Speech and Language Processing via Topic Models
In this tutorial, we will present state-of-art machine learning approaches
for speech and language processing with highlight on the unsupervised methods
for structural learning from the unlabeled sequential patterns. In general,
speech and language processing involves extensive knowledge of statistical models.
We require designing a flexible, scalable and robust system to meet heterogeneous
and nonstationary environments in the era of big data. This tutorial starts from an
introduction of unsupervised speech and language processing based on factor analysis
and independent component analysis. The unsupervised learning is generalized to a latent
variable model which is known as the topic model. The evolution of topic models from
latent semantic analysis to hierarchical Dirichlet process, from non-Bayesian parametric
models to Bayesian nonparametric models, and from single-layer model to hierarchical
tree model shall be surveyed in an organized fashion. The inference approaches based on
variational Bayesian and Gibbs sampling are introduced. We will also present several
case studies on topic modeling for speech and language applications including language model,
document model, retrieval model, segmentation model and summarization model.
At last, we will point out new trends of topic models for speech and language processing.
Organizers: Jen-Tzung Chien
ISCSLP-T4: Deep Learning for Speech Generation and Synthesis
Deep learning, which can represent high-level abstractions in data with an architecture of
multiple non-linear transformation, has made a huge impact on automatic speech recognition (ASR)
research, products and services. However, deep learning for speech generation and synthesis
(i.e., text-to-speech), which is an inverse process of speech recognition (i.e., speech-to-text),
has not generated the similar momentum as it is for ASR yet. Recently, motivated by the success
of Deep Neural Networks in speech recognition, some neural network based research attempts have
been tried successfully on improving the performance of statistical parametric based
speech generation/synthesis. In this tutorial, we focus on deep learning approaches to the
problems in speech generation and synthesis, especially on Text-to-Speech (TTS) synthesis and voice conversion.
First, we give a review for the current main stream of statistical parametric based speech generation
and synthesis, or the GMM-HMM based speech synthesis and GMM-based voice conversion with emphasis
on analyzing the major factors responsible for the quality problems in the GMM-based voice
synthesis/conversion and the intrinsic limitations of a decision-tree based, contextual state
clustering and state-based statistical distribution modeling. We then present the latest deep
learning algorithms for feature parameter trajectory generation, in contrast to deep learning for
recognition or classification. We cover common technologies in Deep Neural Network (DNN) and improved
DNN: Mixture Density Networks (MDN), Recurrent Neural Networks (RNN) with Bidirectional Long Short
Term Memory (BLSTM) and Conditional RBM (CRBM). Finally, we share our research insights and hand-on
experience on building speech generation and synthesis systems based upon deep learning algorithms.
Organizers: Yao Qian and Frank K. Soong |