ISCA - International Speech
Communication Association


ISCApad Archive  »  2015  »  ISCApad #200  »  Journals

ISCApad #200

Friday, February 13, 2015 by Chris Wellekens

7 Journals
7-1Special Issue on 'Signal Processing Techniques for Assisted Listening' of IEEE SIGNAL PROCESSING MAGAZINE
CALL FOR PAPERS
IEEE Signal Processing Society
Special Issue
IEEE SIGNAL PROCESSING MAGAZINE
Special Issue on 'Signal Processing Techniques for Assisted Listening'

 

Aims and Scope
With the rapid advancement in microelectronics and parallel computing, significant computational power is nowadays readily available in ever smaller battery-operated consumer electronics devices. This has paved the way for applications such as active noise cancellation (ANC) headphones, hearing protectors, communication headsets, 3-D glasses, to name a few. In addition, hearing aids have also experienced large advances in electronics and functionality. To a large extent this rapid development can be attributed to the popularity of mobile phones, as these devices are no longer used merely as a communication tool but are multi-media and gaming platforms. Accordingly they require sophisticated processing for augmented reality in which the virtual listening world can be combined conveniently with situational acoustical awareness. The same can be said for assistive listening devices (ALDs), including hearing aids, personal sound amplification devices, and related audio capture accessories. Here the challenge is to render the sound as accessible as possible in order to provide hearing support in challenging acoustical situations. All aforementioned applications are underpinned by fundamental signal processing problems related to sound capture and sound rendering. On the one hand, for sound capture problems such as sensor technology (microphones, accelerometers etc.), acoustic scene analysis, audio signal enhancement, noise suppression with single and multiple sensors, feedback suppression and dereverberation need to be considered. On the other hand, sound rendering involves problems such as active noise cancellation, loudspeaker equalization (for mimicking or adapting outer-ear characteristics), 3-D audio rendering, acoustic scene visualization, automatic mixing and psycho-acoustical processing.

This special issue focuses on technical challenges of assisted listening from a signal processing perspective. Prospective authors are invited to contribute tutorial and survey articles that articulate signal processing methodologies which are critical for applying assisted listening techniques to mobile phones and other communication devices. Of particular interest is the role of signal processing in combining multimedia content, voice communication and voice pick-up in various real-world settings.

Tutorial and survey papers are solicited on advances in signal processing that particularly apply to the following applications:

  • Assistive listening devices, hearing aids and personal sound amplifiers
  • Communication devices
  • Hearing protection and active noise control
  • Navigation systems

These can include the following suggested topics as they relate to the above applications:

Signal processing for robust sound acquisition: Signal processing for acoustic rendering:
Speech enhancement/intelligibility improvement Signal spatialization/3D sound/automatic mixing
Speech separation/separation of non-stationary signals Motion compensation (head tracking, gps systems)
Reverberation reduction Environment-sensitive intelligibility improvement
Array signal processing, and distributed sensors Techniques for natural sound in headphones
Multi-modal acquisition methods  

Submission Process
Articles submitted to this special issue must contain significant relevance to advanced acoustic signal processing enabling assisted listening. All submissions will be peer reviewed according to the IEEE and Signal Processing Society guidelines for both publications. Submitted articles should not have been published or be under review elsewhere. Manuscripts should be submitted online at http://mc.manuscriptcentral.com/sps-ieee using the Manuscript Central interface. Submissions to this special issue of the IEEE SIGNAL PROCESSING MAGAZINE should have significant tutorial value. Prospective authors should consult the site http://www.signalprocessingsociety.org/publications/periodicals/spm/ for guidelines and information on paper submission.

Important Dates: Expected publication date for this special issue is March 2015.

Time Schedule Signal Processing Magazine
White paper (4 pages) due
Invitation notification
Manuscript submission due
Acceptance notification
Revised manuscript due
Final acceptance notification
Final material from authors
Publication date
February 10, 2014
February 24, 2014
May 15, 2014
July 8, 2014
August 20, 2014
September 20, 2014
November 8, 2014 (strict)
March 2015

Guest Editors
Sven Nordholm, Lead GE, Curtin University, Perth, Western Australia (s.nordholm@curtin.edu.au)
Walter Kellermann, Friedrich-Alexander University, Erlangen-Nuremberg, Germany (wk@lnt.de)
Simon Doclo, University of Oldenburg, Oldenburg, Germany (simon.doclo@uni-oldenburg.de)
Vesa Välimäki, Aalto University, Espoo, Finland (vesa.valimaki@alto.fi)
Shoji Makino, University of Tsukuba, Tsukuba, Japan (maki@tara.tsukuba.ac.jp)
John Hershey, Mitsubishi Electric Research Laboratories, Boston, USA (hershey@merl.com)

Back  Top

7-2CSL Special Issue on Speech and Language for Interactive Robots

CSL Special Issue on Speech and Language for Interactive Robots

Aims and Scope

Speech-based communication with robots faces important challenges for their application in real world scenarios. In contrast to conventional interactive systems, a talking robot always needs to take its physical environment into account when communicating with users. This is typically unstructured, dynamic and noisy and raises important challenges. The objective of this special issue is to highlight research that applies speech and language processing to robots that interact with people through speech as the main modality of interaction. For example, a robot may need to communicate with users via distant speech recognition and understanding with constantly changing degrees of noise. Alternatively, the robot may coordinate its verbal turn-taking behaviour with its non-verbal one such as generating speech and gestures at the same time. Speech and language technologies have the potential of equipping robots so that they can interact more naturally with humans, but their effectiveness remains to be demonstrated.  This special issue aims to help fill this gap.

The topics listed below indicate the range of work that is relevant to this special issue, where each article will normally represent one or more topics. In case of doubt about the relevance of your topic, please contact the special issue associate editors.

Topics

  • sound source localization
  • voice activity detection
  • speech recognition and understanding
  • speech emotion recognition
  • speaker and language recognition
  • spoken dialogue management
  • turn-taking in spoken dialogue
  • spoken information retrieval
  • spoken language generation
  • affective speech synthesis
  • multimodal communication
  • evaluation of speech-based human-robot interactions

Special Issue Associate Editors

Heriberto Cuayáhuitl, Heriot-Watt University, UK (contact: hc213@hw.ac.uk)
Kazunori Komatani, Nagoya University, Japan
Gabriel Skantze, KTH Royal Institute of Technology, Sweden

Paper Submission

All manuscripts and any supplementary materials will be submitted through Elsevier Editorial System at http://ees.elsevier.com/csl/. A detailed submission guideline is available as “Guide to Authors” at here. Please select “SI: SL4IR” as Article Type when submitting the manuscripts. For further details or a more in-depth discussion about topics or submission, please contact Guest Editors.

Dates

23 May 2014: Submission of manuscripts
23 August 2014: Notification about decisions on initial submissions
23 October 2014: Submission of revised manuscripts
10 January 2015: Notification about decisions on revised manuscripts
01 March 2015: Submission of manuscripts with final minor changes
31 March 2015: Announcement of the special issue articles on the CSL website http://www.journals.elsevier.com/computer-speech-and-language/

Back  Top

7-3CSL Special Issue on Speech Production in Speech Technologies DEADLINE EXT.

 

CSL Special Issue on Speech Production in Speech Technologies

The use of speech production knowledge and data to enhance speech recognition and other technologies is being actively pursued by a number of widely dispersed research groups using different approaches.  The types of speech production information may include continuous articulatory measurements or discrete-valued articulatory or phonological features.  These quantities might be directly measured, manually labeled, or unobserved but considered to be latent variables in a statistical model.  Applications of production-based ideas include improved speech recognition, silent speech interface, language training tools, and clinical models of speech disorders. 

The goal of this special issue is to highlight the current state of research efforts that use speech production data or knowledge.  The range of data, techniques, and applications currently being explored is growing, and is also benefiting from new ideas in machine learning, making this a particularly exciting time for this research. 

A recent workshop, the 2013 Interspeech satellite workshop on Speech Production in Automatic Speech Recognition (SPASR), as well as the special session on Articulatory Data Acquisition and Processing, brought together a number of researchers in this area.  This special issue will expand on the topics included in these events and beyond. 

Submissions focusing on research in this area are solicited. Topics of interest include, but are not limited to: 

- The collection, labeling, and use of speech production data 

- Acoustic-to-articulatory inversion 

- Speech production models in speech recognition, synthesis, voice conversion, and other technologies 

- Silent speech interfaces 

- Atypical speech production and pathology 

- Articulatory phonology and models of speech production 

  

Submission procedure

Prospective authors should follow the regular guidelines of the Computer Speech and Language Journal for electronic submission (http://ees.elsevier.com/csl). During submission authors must select 'SI: Speech Production in ST' as Article Type. 

 

Review procedure

All manuscripts will be submitted through the editorial submission system and will be reviewed by at least 3 experts.


Schedule:


June 30, 2014: *** New deadline for submissions ***
August 15, 2014:  *** New  Notification of decision ***
September 15, 2014: Deadline for resubmission
November 1, 2014:  Final decision
December 1, 2014: Deadline for camera-ready version
February, 2015:  Publication

 

 
Guest Editors:


Jeff Bilmes, U. Washington, bilmes@uw.edu

Eric Fosler-Lussier, Ohio State U., fosler@cse.ohio-state.edu

Mark Hasegawa-Johnson, U. Illinois at Urbana-Champaign, jhasegaw@uiuc.edu

Karen Livescu, TTI-Chicago, klivescu@ttic.edu

Frank Rudzicz, U. Toronto, frank@cs.toronto.edu


 
Back  Top

7-4Special Issue of ACM Transactions on Accessible Computing (TACCESS) on Speech and Language Interaction for Daily Assistive Technology

Special Issue of ACM Transactions on Accessible Computing (TACCESS) On

Speech and Language Interaction for Daily Assistive Technology

Guest Editors: François Portet, Frank Rudzicz, Jan Alexandersson, Heidi Christensen

Assistive technologies (AT) allow individuals with disabilities to do things that would otherwise be difficult or impossible. Many assistive technologies involve providing universal access, such as modifications to televisions or telephones to make them accessible to those with vision or hearing impairments. An important sub-discipline within this community is Augmentative and Alternative Communication (AAC), which has its focus on communication technologies for those with impairments that interfere with some aspect of human communication, including spoken or written modalities. Another important sub-discipline is Ambient Assisted Living (AAL) which facilitates independent living; these technologies break down the barriers faced by people with physical or cognitive impairments and support their relatives and caregivers. These technologies are expected to improve quality-of-life of users and promote independence, accessibility, learning, and social connectivity.

Speech and natural language processing (NLP) can be used in AT/AAC in a variety of ways including, improving the intelligibility of unintelligible speech, and providing communicative assistance for frail individuals or those with severe motor impairments. The range of applications and technologies in AAL that can rely on speech and NLP technologies is very large, and the number of individuals actively working within these research communities is growing, as evidenced by the successful INTERSPEECH 2013 satellite workshop on Speech and Language Processing for Assistive Technologies (SLPAT). In particular, one of the greatest challenges in AAL is to design smart spaces (e.g., at home, work, hospital) and intelligent companions that anticipate user needs and enable them to interact with and in their daily environment and provide ways to communicate with others. This technology can benefit each of visually-, physically-, speech- or cognitively- impaired persons.

Topics of interest for submission to this special issue include (but are not limited to):

  • Speech, natural language and multimodal interfaces designed for people with physical or cognitive impairments
  • Applications of speech and NLP technology (automatic speech recognition, synthesis, dialogue, natural language generation) for AT applications
  • Novel modeling and machine learning approaches for AT applications
  • Long-term adaptation of speech/NLP based AT system to user's change
  • User studies, overview of speech/NLP technology for AT: understanding the user's needs and future speech and language based technologies.
  • Understanding, modeling and recognition of aged or disordered speech
  • Speech analysis and diagnosis: automatic recognition and detection of speech pathologies and speech capability loss
  • Speech-based distress recognition
  • Automated processing of symbol languages, sign language and nonverbal communication including translation systems.
  • Text and audio processing for improved comprehension and intelligibility, e.g., sentence simplification or text-to-speech
  • Evaluation methodology of systems and components in the lab and in the wild.
  • Resources; corpora and annotation schemes
  • Other topics in AAC, AAL, and AT

 

Submission process

Contributions must not have been previously published or be under consideration for publication elsewhere, although substantial extensions of conference or workshop papers will be considered. as long as they adhere to ACM's minimum standards regarding prior publication (http://www.acm.org/pubs/sim_submissions.html). Studies involving experimentations with real target users will be appreciated. All submissions have to be prepared according to the Guide for Authors as published in the Journal website at http://www.rit.edu/gccis/taccess/

Submissions should follow the journal's suggested writing format (http://www.gccis.rit.edu/taccess/authors.html) and should be submitted through Manuscript Central http://mc.manuscriptcentral.com/taccess , indicating that the paper is intended for the Special Issue. All papers will be subject to the peer review process and final decisions regarding publication will be based on this review.

Important dates:

◦              Extended deadline for full paper submission: 28th April 2014

◦              Response to authors: 30th June 2014

◦              Revised submission deadline: 31st August 2014

◦              Notification of acceptance: 31st October 2014

◦              Final manuscripts due: 30th November 2014

 

 

 

 

 

Back  Top

7-5Revue TAL: numéro spécial sur le traitement automatique du langage parlé (updated)

Premier appel à communications: numéro spécial sur le traitement automatique du langage parlé pour la revue TAL (Traitement Automatique des Langues)

Direction : Laurent Besacier, Wolfang Minker
 
Date limite :  extension au 15 juillet 2014
 
La communication orale reste le moyen le plus naturel pour dialoguer et interagir (avec la machine ou avec une autre personne). Le traitement automatique du langage parlé (TALP) et le dialogue trouvent désormais de nombreuses applications directes dans des domaines divers tels que (liste non exhaustive) la recherche d'information, l'interaction en langue naturelle avec des dispositifs mobiles, la robotique sociale, les technologies d'assistance à la personne, l'apprentissage des langues, etc. Cependant, le TALP pose des problèmes spécifiques liés à la nature même du matériau traité. En effet, on est amené à traiter des  énoncés de parole plus ou moins spontanée et contenant de nombreux traits paralinguistiques. Par exemple, la présence de disfluences orales (répétitions, reprises, incises...) réduit la régularité syntaxique des énoncés ; les énoncés oraux sont également riches d'informations liés aux affects, etc. Par ailleurs, l'étape de transcription automatique, souvent nécessaire avant l'application de traitements de plus haut niveau (compréhension, traduction, analyse, etc.) rend des sorties bruitées (contenant des erreurs) qui nécessitent des analyses robustes et un couplage étroit entre étapes de traitement.  
 

Nous invitons donc les contributions portant sur tout aspect (théorique, méthodologique et pratique) relatif au traitement automatique du langage parlé et à la communication orale, et en particulier (liste non exclusive) : 

  • Reconnaissance automatique de la parole 
  • Compréhension automatique de la parole
  • Traduction de parole
  • Synthèse de la parole
  • Dialogue oral homme - machine 
  • Analyse robuste de la langue parlée
  • Analyse des affects sociaux ou des émotions dans des énoncés oraux
  • Fouille de documents à composante orale
  • Applications à composantes orales (recherche d'information, interaction, robotique, etc)
  • Outils d'aide à l'apprentissage d’une langue seconde
  • Aspects multilingues du traitement automatique du langage parlé
  • Evaluation de systèmes de traitement du langage parlé
  • Corpus et ressources pour l'oral
  • Analyse du discours oral
  • Dialogue adaptatif au contexte et au profil de l'utilisateur
  • Analyse des traits paralinguistiques dans des énoncés oraux 

 

ÉDITEURS INVITÉS 

Laurent Besacier 
Wolfang Minker

 

COMITE SCIENTIFIQUE

Une version provisoire (à confirmer) des membres est disponible sur: http://tal-55-2.sciencesconf.org/resource/page/id/2 

 

LANGUE
Les articles sont écrits en français ou en anglais. Les soumissions en anglais ne sont acceptées que pour les auteurs non francophones.


FORMAT DE LA SOUMISSION

Les articles doivent être déposés sur la plateforme http://tal-55-2.sciencesconf.org/

La revue ne publie que des contributions originales, en français ou en anglais.

Les papiers acceptés feront au maximum 25 pages en PDF. Le style est disponible pour téléchargement sur le site du journal TAL 

 

CONTACT

Laurent Besacier     (Laurent.Besacier@imag.fr)

Wolfgang Minker      (Wolfgang.Minker@uni-ulm.de)

 
Back  Top

7-6IEEE Journal of Selected Topics in Signal Processing: Special Issue on Spatial Audio
CALL FOR PAPERS
IEEE Journal of Selected Topics in Signal Processing
Special Issue on Spatial Audio

 

Spatial audio is an area that has gained in popularity in the recent years. Audio reproduction setups evolved from the traditional two-channel loudspeaker setups towards multi-channel loudspeaker setups. Advances in acoustic signal processing even made it possible to create a surround sound listening experience using traditional stereo speakers and headphones. Finally, there has been an increased interest in creating different sound zones in the same acoustic space (also referred to as personal audio). At the same time, the computational capacity provided by mobile audio playback devices has increased significantly. These developments enable new possibilities for advanced audio signal processing, such that in the future we can record, transmit and reproduce spatial audio in ways that have not been possible before. In addition, there have been fundamental advances in our understanding of 3D audio.

Due to the increasing number of different formats and reproduction systems for spatial audio, ranging from headphones to 22.2 speaker systems, it is major challenge to ensure interoperability between formats and systems, and consistent delivery of high-quality spatial audio. Therefore, the MPEG committee is in the process of establishing new standards for 3D Audio Content Delivery.

The scope of this Special Issue on Spatial Audio is open to contributions ranging from the measurement and modeling of an acoustic space to reproduction and perception of spatial audio. While individual submissions may focus on any of the sub-topics listed below, papers describing a larger spatial audio signal processing systems will be considered as well.

We invite authors to address some of the following spatial audio aspects:

    • Capture of Spatial Sound, use of different microphone arrays to record 3D sound fields
    • Loudspeaker and Headphone Reproduction of Spatial Sound, including e.g. wave field synthesis, Ambisonics, arbitrary multi-channel loudspeaker setups, cross-talk cancellation systems, and personal audio systems
    • Spatial Sound Processing including e.g. downmixing, upmixing, spatial sound enhancement, and reverberation effects
    • Sound Source Localization and Room Geometry Estimation, advanced analysis of audio signals for reconstruction of the acoustic environment
    • Room Acoustics Modeling covering all different modeling techniques ranging from computationally heavy wave-based techniques and geometrical acoustics to lightweight perceptually-based models.

Prospective authors should visit http://www.signalprocessingsociety.org/publications/periodicals/jstsp/ for information on paper submission. Manuscripts should be submitted at http://mc.manuscriptcentral.com/jstsp-ieee.

Manuscript Submission: July 1, 2014
First Review Due: October 15, 2014
Revised Manuscript: December 1, 2014
Second Review Due: February 1, 2015
Final Manuscript: March 1, 2015

Guest Editors:
Lauri Savioja, Aalto University, Finland (Lauri.Savioja@aalto.fi)
Akio Ando, University of Toyama, Japan (andio@eng.u-toyama.ac.jp)
Ramani Duraiswami, University of Maryland, USA (ramani@umiacs.umd.edu)
Emanuël Habets, Int. Audio Laboratories Erlangen, Germany (emanuel.habets@audiolabs-erlangen.de)
Sascha Spors, Universität Rostock, Germany (sascha.spors@uni-rostock.de)

Back  Top

7-7Revue TAL: numéro spécial numéro spécial sur le traitement automatique du langage parlé

 

Special issue on spoken language processing

Guest editors: Laurent Besacier, Wolfgang Minker

 

Speech is the most natural way to communicate and interact (with the machine or with another person) . Spoken language processing and dialogue have now many direct applications in various areas such as (but not limited to) information retrieval, natural language interaction with mobile devices, social robotics, assistive technologies, technologies for language learning, etc. . However, spoken language processing poses specific problems related to the nature of the speech material itself. Indeed, spontaneous speech utterances have to be processed and they contain many paralinguistic features. For instance, disfluencies (repetitions , false starts, etc.) reduces the syntactic regularity of utterances. Moreover, spontaneous utterances convey rich information related to emotions , etc. Furthermore, automatic speech recognition (ASR) step, often required before the application of higher level processing (understanding , translation, analysis, etc.), produces noisy outputs (with errors ) which require robust and tight coupling between modules.

 We invite contributions on any aspect (theoretical, methodological and practical) of spoken language processing and oral communication ; in particular (non-exclusive list):

 -Automatic speech recognition

-Spoken language understanding

-Speech translation

-Text-to-Speech synthesis

-Man-machine dialogue

-Robust analysis of spoken language

-Analysis of social affects or emotions in spontaneous speech

-Mining spoken language documents

-Spoken language applications (mobile interaction, robotics, etc. )

-Technologies for language learning

-Multilingual aspects of spoken language processing

-Evaluation for spoken language processing

-Corpora and resources for spoken language

-(Spoken) discourse analysis

-Adaptive dialogue (context, user profile)

-Analysis of paralinguistic features in spoken language

 

IMPORTANT DATES

-call : march 2014
-submission of contributions : 30 june 2014
-first authors notification : 15 september 2014 
-publication : end 2014 / begin 2015

Submission format

LANGUAGE
Manuscripts may be submitted in English or French. French-speaking authors are requested to submit their contributions in French.

PAPER SUBMISSION
Papers must describe original, completed, and unpublished work.  Each submission will be reviewed by two programme committee members. 
Papers must be submitted on Sciencesconf platform  http://tal-55-2.sciencesconf.org/ 

Accepted papers will be maximum 25 pages long in PDF. Style sheets are available for download on the Web site of the TAL journal

Back  Top

7-8Journal of Natural Language Engineering - Special Issue on “Machine Translation Using Comparable Corpora”

 

***** Journal of Natural Language Engineering - Special Issue on “Machine Translation Using Comparable Corpora” *****

CALL FOR PAPERS Statistical machine translation based on parallel corpora has been very successful. The major search engines' translation systems, which are used by millions of people, are primarily using this approach, and it has been possible to come up with new language pairs in a fraction of the time that would be required when using more traditional rule-based methods. In contrast, research on comparable corpora is still at an earlier stage. Comparable corpora can be defined as monolingual corpora covering roughly the same subject area in different languages but without being exact translations of each other. However, despite its tremendous success, the use of parallel corpora in MT has a number of drawbacks: 1) It has been shown that translated language is somewhat different from original language, for example Klebanov & Flor showed that 'associative texture' is lost in translation. 2) As they require translation, parallel corpora will always be a far scarcer resource than comparable corpora. This is a severe drawback for a number of reasons: a) Among the about 7000 world languages, of which 600 have a written form, the vast majority are of the 'low resource' type. b) The number of possible language pairs increases with the square of the number of languages. When using parallel corpora, one bitext is needed for each language pair. When using comparable corpora, one monolingual corpus per language suffices. c) For improved translation quality, translation systems specialized on particular genres and domains are desirable. But it is far more difficult to acquire appropriate parallel rather than comparable training corpora. d) As language evolves over time, the training corpora should be updated on a regular basis. Again, this is more difficult in the parallel case. For such reasons it would be a big step forward if it were possible to base statistical machine translation on comparable rather than on parallel corpora: The acquisition of training data would be far easier, and the unnatural 'translation bias' (source language shining through) within the training data could be avoided. But is there any evidence that this is possible? Motivation for using comparable corpora in MT research comes from a cognitive perspective: Experience tells that persons who have learned a second language completely independently from their mother tongue can nevertheless translate between the languages. That is, human performance shows that there must be a way to bridge the gap between languages which does not rely on parallel data. Using parallel data for MT is of course a nice shortcut. But avoiding this shortcut by doing MT based on comparable corpora may well be a key to a better understanding of human translation, and to better MT quality. Work on comparable corpora in the context of MT has been ongoing for almost 20 years. It has turned out that this is a very hard problem to solve, but as it is among the grand challenges in multilingual NLP, interest has steadily increased. Apart from the increase in publications this can be seen from the considerable number of research projects (such as ACCURAT and TTC) which are fully or partially devoted to MT using comparable corpora. Given also the success of the workshop series on “Building and Using Comparable Corpora“ (BUCC), which is now in its seventh year, and following the publication of a related book (http://www.springer.com/computer/ai/book/978-3-642-20127-1), we think that it is now time to devote a journal special issue to this field. It is meant to bundle the latest top class research, make it available to everybody working in the field, and at the same time give an overview on the state of the art to all interested researchers.

TOPICS OF INTEREST We solicit contributions including but not limited to the following topics: • Comparable corpora based MT systems (CCMTs) • Architectures for CCMTs • CCMTs for less-resourced languages • CCMTs for less-resourced domains • CCMTs dealing with morphologically rich languages • CCMTs for spoken translation • Applications of CCMTs • CCMT evaluation • Open source CCMT systems • Hybrid systems combining SMT and CCMT • Hybrid systems combining rule-based MT and CCMT • Enhancing phrase-based SMT using comparable corpora • Expanding phrase tables using comparable corpora • Comparable corpora based processing tools/kits for MT • Methods for mining comparable corpora from the Web • Applying Harris' distributional hypothesis to comparable corpora • Induction of morphological, grammatical, and translation rules from comparable corpora • Machine learning techniques using comparable corpora • Parallel corpora vs. pairs of non-parallel monolingual corpora • Extraction of parallel segments or paraphrases from comparable corpora • Extraction of bilingual and multilingual translations of single words and multi-word expressions, proper names, and named entities from comparable corpora

IMPORTANT DATES December 1, 2014: Paper submission deadline February 1, 2015: Notification May 1, 2015: Deadline for revised papers July 1, 2015: Final notification September 1, 2015: Final paper due

GUEST EDITORS Reinhard Rapp, Universities of Aix Marseille (France) and Mainz (Germany) Serge Sharoff, University of Leeds (UK) Pierre Zweigenbaum, LIMSI, CNRS (France) FURTHER INFORMATION Please use the following e-mail address to contact the guest editors: jnle.bucc (at) limsi (dot) fr Further details on paper submission will be made available in due course at the BUCC website: http://comparable.limsi.fr/bucc2014/bucc-introduction.html

Back  Top

7-9IEEE/ZACM Trans. ASLP: special issue on continuous space and related methods in natural language processing
CALL FOR PAPERS
IEEE/ACM Transactions on Audio, Speech and Language Processing
Special issue on continuous space and related methods in natural language processing

 

Natural Language Processing (NLP) aims to analyze, understand, and generate languages that humans use naturally. Significant progress in NLP has been achieved in recent years, addressing important and practical real - world problems , enabling mass deploymen t of large - scale systems . New machine learning paradigms such as deep learning and continuous space methods have contributed to inferring language patterns from increasingly large real - world data and to making predictions about new data more accurate.

One of the challenges in NLP is to represent language in a form that can be processed effectively by computing algorithms. Words in sequence are traditionally treated as discrete symbols, which ha s its advantages and limitations. The research on continuous space methods provides a promising alternative that describes words and their semantic and syntactic relationships in a different way. In continuous space language modeling , we represent words with real - valued vectors. In this way, conditional probability distributions of words can be learn ed and expressed as smooth function s of these vectors; similar words are therefore described as neighbors in a continuous space. A Neural Network Language Model is a typical example of such continuous space method s .

Building on the success of acoustic and statistical language modeling , research on artificial (deep) neural networks and continuous space models in general has seen significant progress in mitigating data sparseness , incorporating longer contexts, and mode ling morphological, syntactic and semantic relationships across words. As a result, continuous space models are now embedded in many state - of - the - art speech recognition and machine translation systems . This special issue provides a forum to discuss the lat est findings on research problems related to the application of continuous space and related models in NLP. We invite papers on various NLP topics, including but not limited to:

- Automatic speech recognition
- Speaker recognition
- Language modeling
- Machine translation
- Spoken language understanding
- Spoken document retrieval
- Text Mining
- Computational semantics
- Morphological Analysis
- Syntactic Parsing
- Discourse and Dialogue
- Machine learning methods

The authors are required to follow the Author’s Guide for manuscript submission to the IEEE /ACM Transactions on Audio, Speech, and Language Processing at http://www.signalprocessingsociety.org/publications/periodicals/taslp/taslp-author-information

- Submission deadline: July 27, 2014
- Notification of first round review: Oct. 1 , 2014
- Notification of acceptance: Dec. 1 , 201 4
- Final manuscripts due: Dec. 15 , 2015

For further information, please contact the guest editors:

- Haizhou Li, Institute for Infocomm Research, Singapore
- Marcello Federico, FBK, Italy
- Xiaodong He, Microsoft Research, USA
- Helen Meng, The Chinese University of Hong Kong, China
- Isabel Trancoso, INESC-ID and Instituto Superior Técnico, University of Lisbon, Portugal

Copyright IEEE Signal Processing Society. All rights reserved.
SPS Home | IEEE







This email was sent to christian.wellekens@eurecom.fr
why did I get this?    unsubscribe from this list    update subscription preferences
IEEE SPS · 445 Hoes Lane · Piscataway, NJ 08854 · USA

Back  Top

7-10CfP Special issue Speech Communication on Advances in Sparse Modeling and Low-rank Modeling for Speech Processing
Manuscript due: Dec. 1, 2014

Journal: Speech Communication

Description:

Sparse and low-rank modeling aim to incorporate the low-dimensional
structures pertained to the geometry of the underlying problems to
achieve the optimal solution. These concepts have been proven to be very
effective for a wide range of applications at the intersection of
multiple fields, including machine learning, signal processing and
statistics.

In the context of audio and speech processing, and more particularly
multiparty communications in reverberant and overlapping conditions, the
integration of sparse and low-rank modeling concepts has lead to several
interesting new directions and promising results in speech communication
problems, ranging from denoising to deconvolution and from separation to
recognition. Several other exciting developments include sparse linear
prediction, missing data recovery, audio content analysis and
inpainting. Addressing such real applications is particularly
challenging due to the complex acoustic and speech characteristics, and
the need to develop new modeling strategies that meet the foundational
theoretical hypotheses. In addition, speech recognition performance
seems to degrade in these complex acoustic conditions, and thus research
in this direction is critical from both a theoretical and industry
perspective.

The goal of the proposed special issue is to consolidate the research in
these diverse fields in a coherent framework and overview the recent
advances and trends where sparse and low-rank modeling and applications
are converging to new fundamental and practical paradigms that could
also lead to the emergence of new speech technologies.



Topics of interest include:

* Manifold learning in speech processing: single and multi-microphone
speech enhancement
* Sparse modeling and low-rank modeling for separation and denoising
* Sparse regression and classification
* Sparse dimensionality reduction for feature extraction
* Structured sparsity models underlying audio and speech representation
* Auditory-inspired sparse modeling
* Sparse modeling and low-rank modeling for source localization
* Sparse representation and low-rank representation for reverberant
acoustic modeling
* Sparse data processing and modeling in low-resourced languages
* Applications in speech recognition, privacy-preserving speech
processing, speaker recognition and authentication, speaker diarization,
microphone array calibration, audio information retrieval, speech
synthesis and coding

Lead Guest Editors:

* Prof. Hervé Bourlard, herve.bourlard@idiap.ch
* Dr. Afsaneh Asaei, afsaneh.asaei@idiap.ch
* Dr. Tara N. Sainath, tsainath@us.ibm.com
* Prof. Sharon Gannot, sharon.gannot@biu.ac.il

For more information about this special issue, please visit:

http://si.eurasip.org/issues/36/advances-in-sparse-modeling-and-low-rank-modeling/
_______________________________________________
Back  Top

7-11[Special Issue] Deep Learning for Speech and Language Processing Applications, EURASIP
Manuscript due: Dec. 15, 2014

Journal: EURASIP Journal on Audio, Speech, and Music Processing

Description:

Deep learning techniques have enjoyed enormous success in the speech and
language processing community over the past few years, beating previous
state-of-the-art approaches to acoustic modeling, language modeling, and
natural language processing. A common theme across different tasks is
that that the depth of the network allows useful representations to be
learned. For example, in acoustic modeling, the ability of deep
architectures to disentangle multiple factors of variation in the input,
such as various speaker-dependent effects on speech acoustics, has led
to excellent improvements in speech recognition performance on a wide
variety of tasks. In addition, in natural language processing and
language modeling tasks, integrating learned vector space models of
words, which perform smoothing and clustering based on semantic and
syntactic information contained in word contexts, with recurrent or
recursive architectures has led to significant advances.
 
We as a community should continue to understand what makes deep learning
successful for speech and language, and how further improvements can be
achieved. For example, just as deep networks made us re-think the input
feature representation pipeline used for speech recognition, we should
continue to push deep learning into other areas of the speech
recognition pipeline.
 
In addition, new architectures, such as convolutional neural networks
and recurrent networks using long short-term memory cells, have improved
performance, and we believe alternative architectures can improve
performance further. Secondly, optimization of large neural network
models remains a huge challenge, both because of computational cost and
amount of data, which could possibly be unsupervised.


Topics of interest include:

* New deep-learning architectures and algorithms
* Optimization strategies for deep learning
* Improved adaptation methods for deep learning
* Unsupervised and semi-supervised training for deep learning
* Novel applications of deep learning for speech and language tasks
* Theoretical and empirical understanding of deep learning for speech and 
 language
* Deep-learning toolkits and/or platforms for big data

Lead Guest Editor:

* Tara Sainath, Google Inc., USA

Guest Editors:

* Michiel Bacchiani, Google Inc., USA
* Hui Jiang, York University, Canada
* Brian Kingsbury, Thomas J. Watson Research Center, USA
* Hermann Ney, RWTH Aachen, Germany
* Frank Seide, Microsoft Research Asia, China
* Andrew Senior, Google Inc., USA

For more information about this special issue, please visit:

http://si.eurasip.org/issues/38/deep-learning-for-speech-and-language-processing/
_______________________________________________
Back  Top

7-12SPECIAL ISSUE ON DIALOGUE STATE TRACKING
CALL FOR PAPERS  
DIALOGUE & DISCOURSE 
SPECIAL ISSUE ON DIALOGUE STATE TRACKING 

GUEST EDITORS 

Jason D. Williams, Microsoft Research
Antoine Raux, Lenovo
Matthew Henderson, Cambridge University

IMPORTANT DATES

Submission deadline:                    3 April 2015
Notification:                           26 June 2015
Final version of accepted papers due:   28 August 2015
Anticipated publication:                16 October 2015

INTRODUCTION 

Conversational systems are increasingly becoming a part of daily life, with examples including Apple's Siri, Google Now, 
Nuance Dragon Go, Xbox and Cortana from Microsoft, and numerous new entrants.  Many conversational systems include 
a dialogue state tracking function, which estimates relevant aspects of the interaction such as the user's goal, level of 
frustration, trust towards the system, etc, given all of the dialogue history so far.  For example, in a tourist information system,
 the dialogue state might indicate the type of business the user is searching for (pub, restaurant, coffee shop), their desired 
price range and type of food served.  Dialogue state tracking is difficult because automatic speech recognition (ASR) and 
spoken language understanding (SLU) errors are common, and can cause the system to misunderstand the user.  
At the same time, state tracking is crucial because the system relies on the estimated dialogue state to choose actions -- for 
example, which restaurants to suggest.

Most commercial systems use hand-crafted heuristics for state tracking, selecting the SLU result with the highest confidence score,
 and discarding alternatives. In contrast, statistical approaches consider many hypotheses for the dialogue state.  By exploiting 
correlations between turns and information from external data sources -- such as maps, knowledge bases, or models of past 
dialogues -- statistical approaches can overcome some SLU errors.

Although dialogue state tracking has been an active area of study for more than a decade, there has been a flurry of new work 
in the past 2 years.  This has been driven in part by the availability of common corpora and evaluation measures provided by a 
series of three research community challenge tasks called the Dialogue State Tracking Challenge.  With these resources, 
researchers are able to study dialogue state tracking without investing the time and effort required to build and operate a spoken 
dialogue system.  Shared resources also allow direct comparison of methods across research groups.  Results from the Dialogue 
State Tracking Challenge have been presented at special sessions in SIGDIAL 2013, SIGDIAL 2014, and IEEE SLT 2014.

TOPICS OF INTEREST

The aim of this special issue is to provide a forum for in-depth, journal-level work on dialogue state tracking.  

This issue welcomes papers covering any topic relevant to dialogue state tracking.  Specific examples include (but are not
 limited to):

- Algorithms for dialogue state tracking, including those based on machine learning or novel heuristics
- Adaptation and learning in dialogue state tracking, for example across domains, users, usage environments, etc.
- Analyses of dialogue state tracking methods, or analyses of characteristics of dialogue that affect dialogue state tracking
- Investigations of metrics used for dialogue state tracking, including the impact of dialogue state tracking on end-to-end 
dialogue systems
- Descriptions and analyses of resources for dialogue state tracking, including corpora 
- Applications of dialogue state tracking to new domains or new settings, such as multi-modal systems

Submissions should report on new work, or substantially expand on previously published work with additional experiments, 
analysis, or important detail.  Previously-published aspects may be included but should be clearly indicated.

RELEVANT RESOURCES

All data from the dialogue state tracking challenge series continues to be available for use, including the dialogue data itself,
 scripts for evaluation and baseline trackers, raw output from trackers entered in the challenges, and performance summaries. 
If your work is on dialogue state tracking for information-seeking dialogues and/or you think the data is appropriate, you are 
strongly encouraged to report results on these data, to enable comparison.

The dialogue state tracking challenge data is available here:

- Dialogue State Tracking Challenge 1:    http://research.microsoft.com/en-us/events/dstc/
- Dialogue State Tracking Challenge 2&3:  http://camdial.org/~mh521/dstc/

SUBMISSIONS

Papers should be submitted on the Dialogue & Discourse journal website, following instructions and formatting guidelines given
 there:

http://www.dialogue-and-discourse.org/submission.shtml

Submitted papers will be reviewed according to the Dialogue & Discourse reviewing criteria and appropriateness to the topic of 
the special issue.

CONTACT

Contact Jason Williams (jason.williams@microsoft.com) for further information about this call for papers.
Back  Top

7-13Special issue of Eurasip Journal in Adv.Signal Proc. 'Silencing the Echoes' - Processing of Reverberant Speech

'Silencing the Echoes' - Processing of Reverberant Speech

Submission Instructions

asp.eurasipjournals.com

Manuscript Due

Feb. 1, 2015 (in 3 months, 3 weeks)

Description

The reverberation contained in speech signals captured by distant microphones reduces both the perceptual speech quality and the performance of automatic speech recognition (ASR) systems, hampering both human-human communication and human-machine interaction. Thus, for applications that depend on distant sound capture, processing of the speech signal to mitigate the reverberation problem is essential. Such applications include voice control of consumer products, e.g. smart phones, wearable devices, interactive TVs, and appliances in smart homes; hearing aids; voice communication with robots and avatars; automatic meeting transcription; speech recognition in call centers; automatic annotation of videos; speech-to-speech translation.

To enable researchers in the field of reverberant speech processing to carry out comprehensive evaluations of their methods based on a common database and common evaluation metrics, recently the Reverberant Voice Enhancement and Recognition Benchmark (REVERB) challenge has been organized. Inspired by the great interest induced by this challenge, we invite contributions on processing of reverberant speech signals both for signal enhancement to increase perceptual speech quality and for robust recognition of reverberant speech.

We strongly encourage the evaluation of the proposed approaches on the data provided by the REVERB challenge. Alternatively or additionally, well-documented and publicly available, databases can be used. We invite challenge participants to submit extended descriptions of their (enhanced) systems. Participation in the challenge is however not a prerequisite for submitting a contribution to this special issue.

Topics of interest include:
  • Single-channel and multichannel speech dereverberation algorithms
  • ASR robust to reverberation
  • Interconnection of dereverberation and ASR systems
  • Objective measures for evaluating the perceptual speech quality of dereverberated signals
  • Estimation of reverberation parameters (e.g. reverberation time, direct-to-reverberation ratio)
  • Dereverberation in dynamic scenarios
  • Highly reverberant scenarios (reverberation times exceeding one second, e.g., concert halls, museums, terminal halls, houses of worship)
  • System design for handling reverberant speech

Guest Editors

  • Sharon Gannot, Faculty of Engineering, Bar-Ilan University, Israel
  • Armin Sehr, Beuth Hochschule für Technik Berlin
  • Emanuël Habets, Friedrich-Alexander-University of Erlangen-Nuremberg
  • Keisuke Kinoshita, NTT Communication Science Laboratories
  • Walter Kellermann, University Erlangen-Nuernberg, Germany
  • Reinhold Haeb-Umbach, Department of Communications Engineering, University of Paderborn, Germany
EURASIP Journal on Advances in Signal Processing

Back  Top

7-14Speech Communication Special issue on Advances in Sparse Modeling and Low-rank Modeling for Speech Processing

Manuscript due: Jan. 10, 2015

Journal: Speech Communication

Description:

Sparse and low-rank modeling aim to incorporate the low-dimensional
structures pertained to the geometry of the underlying problems to
achieve the optimal solution. These concepts have been proven to be very
effective for a wide range of applications at the intersection of
multiple fields, including machine learning, signal processing and
statistics.

In the context of audio and speech processing, and more particularly
multiparty communications in reverberant and overlapping conditions, the
integration of sparse and low-rank modeling concepts has lead to several
interesting new directions and promising results in speech communication
problems, ranging from denoising to deconvolution and from separation to
recognition. Several other exciting developments include sparse linear
prediction, missing data recovery, audio content analysis and
inpainting. Addressing such real applications is particularly
challenging due to the complex acoustic and speech characteristics, and
the need to develop new modeling strategies that meet the foundational
theoretical hypotheses. In addition, speech recognition performance
seems to degrade in these complex acoustic conditions, and thus research
in this direction is critical from both a theoretical and industry
perspective.

The goal of the proposed special issue is to consolidate the research in
these diverse fields in a coherent framework and overview the recent
advances and trends where sparse and low-rank modeling and applications
are converging to new fundamental and practical paradigms that could
also lead to the emergence of new speech technologies.


Topics of interest include:

* Manifold learning in speech processing: single and multi-microphone
speech enhancement
* Sparse modeling and low-rank modeling for separation and denoising
* Sparse regression and classification
* Sparse dimensionality reduction for feature extraction
* Structured sparsity models underlying audio and speech representation
* Auditory-inspired sparse modeling
* Sparse modeling and low-rank modeling for source localization
* Sparse representation and low-rank representation for reverberant
acoustic modeling
* Sparse data processing and modeling in low-resourced languages
* Applications in speech recognition, privacy-preserving speech
processing, speaker recognition and authentication, speaker diarization,
microphone array calibration, audio information retrieval, speech
synthesis and coding

Guest Editors:

* Prof. Hervé Bourlard, herve.bourlard@idiap.ch
* Dr. Afsaneh Asaei, afsaneh.asaei@idiap.ch
* Dr. Tara N. Sainath, tsainath@google.com
* Prof. Sharon Gannot, sharon.gannot@biu.ac.il

For more information about this special issue, please visit:

http://si.eurasip.org/issues/48/advances-in-sparse-modeling-and-low-rank-modeling/

Back  Top

7-15Special issue of Speech Communication on 'Phase-Aware Signal Processing

Special issue of Speech Communication on 'Phase-Aware Signal Processing in
Speech Communication'
http://si.eurasip.org/issues/46/phase-aware-signal-processing-in-speech/

Back  Top

7-16JAIR Special Track on Cross-language Algorithms and Applications

JAIR Special Track on Cross-language Algorithms and Applications


Track Editor
Lluís Màrquez, Qatar Computing Research Institute

Associate Track Editors
Marta R. Costa­jussà, Instituto Politécnico Nacional
Srinivas Bangalore, AT&T Labs-Research
Patrik Lambert, Universitat Pompeu Fabra
Elena Montiel-Ponsoda, Universidad Politécnica de Madrid


The Journal of Artificial Intelligence Research (JAIR) is pleased to
announce the launch of the Special Track on Cross-language Algorithms and
Applications. The core Artificial Intelligence technologies of speech and
natural language processing need to address the challenges of processing
multiple languages. While the first challenge of multilingualism is to
bridge the nomenclature gap for the same concepts, the next significant
challenge is to develop algorithms and applications that not only scale to
multiple languages but also leverage cross-lingual similarities for
improved natural language processing.

The goal of this special track is to serve as a home for the publication of
leading research on Cross-language Algorithms and Applications, focusing on
developing unified themes leading to the development of the science of
multi- and cross-lingualism.  Topics of interest include, but are not
limited to: efforts in the direction of multilingual transliteration;
multilingual document summarization; rapid prototyping of cross language
tools for low resource languages; and machine translation.

Articles published in the Cross-language Algorithms and Applications track
must meet the highest quality standards as measured by originality and
significance of the contribution and clarity of presentation. Papers will
be coordinated by the track editor and associate editors, and reviewed by
peer reviewers drawn from the JAIR Editorial Board and the larger
community. All articles should be submitted using the normal JAIR
submission process. Please indicate that the submission is intended for the
Special Track in the section 'Special Information for editors'.


For more information and submission instructions, please see:
http://www.jair.org/specialtrack-claa.html


Timetable

1st March 2015              Deadline for Submissions
1st June 2015                 Notification of Acceptance/Revision/Rejection
15th July 2015                Deadline for Re-submission of papers requiring revision
15th September 2015    Notification of Final Acceptance
1st November 2015        Final manuscript due

Contact: martaruizcostajussa@gmail.com

Submission Instructions: Use JAIR conventional submissions instructions
available at http://www.jair.org/submission_info.html

Back  Top

7-17Neew journal: Advances in Applied Acoustics

Advances in Applied Acoustics

ticsCall for papers

Advances in Applied Acoustics (AIAAS) is an open-access and peer-reviewed journal. The journal publishes original papers at the forefront of the latest advancements in applied acoustics. The contributions are published in both printed and online version.

Your paper can be published with no charge if accepted.


Indexing

Google
Cloud D
JournalTOCs
Bible
WorldCat
INK ER
Deepdyve
getCITED
dogpile
pubzone
crossref
Academia.edu
AoL.
Electronic Journals Library
Yandex
Scribd
GitHub
CNKI SCHOLAR
ARCHIVE


Scope

-Environmental Noise
-Audiology Studies
-Occupational Noise Management
-Physiology of Hearing
-Engineering Noise Control
-Application of Acoustics to Medicine
-Innovative Applications of Materials for
Acoustic Purposes
-Detection and localization of Marine
Mamals Using Passive Acoustics
-Musical Acoustics
-Sound Shielding in The Presence of
Turbulence
-Marine Acoustics
-Architectural Acoustics
-Time-domain Modelling in Outdoor Sound Propagation
-Urban Acoustics
-Head- Related Transfer Function and its Applications
More...


Get more information by contacting aiaa@seipub.org or visiting the website.

Back  Top



 Organisation  Events   Membership   Help 
 > Board  > Interspeech  > Join - renew  > Sitemap
 > Legal documents  > Workshops  > Membership directory  > Contact
 > Logos      > FAQ
       > Privacy policy

© Copyright 2024 - ISCA International Speech Communication Association - All right reserved.

Powered by ISCA