ISCA - International Speech
Communication Association


ISCApad Archive  »  2024  »  ISCApad #316  »  Academic and Industry Notes

ISCApad #316

Thursday, October 10, 2024 by Chris Wellekens

4 Academic and Industry Notes
4-1New Master curriculum integrating advanced study and research covering all areas of language science,Univ. of Paris, France

The Paris Graduate School of Linguistics (PGSL) is a newly-formed Paris-area graduate program covering all areas of language science.

It offers a comprehensive Master curriculum integrating advanced study and research, in close connection with PhD programs as well as with the Empirical Foundations of Linguistics consortium. 

Research plays a central part in the program, and students also take elective courses to develop an interdisciplinary outlook. Prior knowledge of French is not required.

For more details, please see https://paris-gsl.org/index.html

New funding opportunity: https://u-paris.fr/en/call-for-applications-international-students-miem-scholarship-program/

 Application deadline : February 1st 2021 (program starting September 1st 2021)

PGSL is funded by Smarts-UP (Student-centered iMproved, Active Research-based Training Strategy at Université de Paris) through the ANR SFRI grant « Grandes universités de recherche » (PIA3) 2020-2029.

Back  Top

4-2Cambridge's Machine Learning and Machine Intelligence MPhil

Cambridge's Machine Learning and Machine Intelligence MPhil

 

Are you interested in speech and language processing, computer vision & robotics, human-computer interaction, or machine learning? Please consider applying to the University of Cambridge?s Machine Learning and Machine Intelligence (MLMI) MPhil programme

 

The MLMI MPhil is an elite 11 month programme with a small cohort of about 30 students each year. Due to its small size there is the opportunity to carry out PhD-like research projects on the course (see here for previous students? dissertations), as well having a number of bespoke taught modules with lots of opportunities to interact with the faculty and other members of the course (see here for a list of modules and here for a list of the teaching staff).

 

Previous members of the MPhil have gone on to study for PhDs in top research groups (e.g. at Oxford, Cambridge, Stanford, and MIT), and have gone into top industry positions (e.g. Microsoft Research, Facebook AI Research, Open AI, and AstraZeneca).

 

This year our programme is restructuring around four overlapping tracks: speech and language processing, computer vision & robotics, human-computer interaction, and machine learning. You apply to one of these tracks and this choice shapes your module options and the research project that you will take on. We are especially interested in candidates who are interested in speech and language processing, computer vision & robotics, and human-computer interaction as we have significant capacity to expand in these areas this year.

 

Details about the application process can be found on our website. The application deadline is 2nd December 2021.

Back  Top

4-3Serveur Discord pour jeunes chercheurs.
Nous souhaitons vous faire part aujourd'hui de la création d'un serveur Discord pour les jeunes chercheurs en parole: https://discord.gg/kSgaZp7yg9

Cet espace de discussion a pour but de rassembler la communauté des jeunes chercheurs en parole en France (étudiants en master, doctorants, post-doctorants...).
Vous pourrez notamment y partager vos derniers papiers publiés, vos questions ou appels à l'aide pour l'utilisation d'un logiciel (ou autre), ou simplement rentrer en contact avec des personnes qui travaillent dans votre domaine ou des domaines liés au votre. Rien n'est fixe, le serveur est voué à évoluer au cours de son utilisation! Il peut aussi nous servir à nous retrouver lors d'une conférence ou autre. En tant que jeunes chercheurs avec un petit réseau, on connaît tous ce sentiment peu confortable d'être seul(e) à une conférence, malgré les 1500 personnes autour de nous. Alors plutôt que de stresser chacun dans notre coin, autant nous retrouver et partager tout ça ensemble en nous donnant rendez-vous grâce au serveur de discussion!

La création de ce serveur Discord fait suite à l'appel du comité d'organisation des JEP pour l'organisation d'un événement à destination des jeunes chercheurs en parole lors des Journées d'Etudes en Parole qui auront lieu à Noirmoutier du 13 au 17 juin 2022 (https://jep2022.univ-nantes.fr/). Nous l'avons d'abord créé pour discuter entre jeunes chercheurs sur ce que l'on pourrait imaginer comme événement, puis nous avons pensé qu'il serait intéressant de l'ouvrir à tous pour recueillir vos besoins/envies et voir ce qui pourrait intéresser un maximum de personnes. Il est donc également destiné à essayer de mieux cerner les besoins de formation de chacun. Rassurez-vous, il est partagé en deux: une catégorie pour tous, et une catégorie pour celles et ceux intéressés par l'organisation de journées d'études, ainsi, pas de spam inutile. Si vous souhaitez rejoindre cette deuxième catégorie, il faudra me le notifier dans le canal #général ou me le demander par message privé*.

Rendez-vous sur le serveur pour faire vivre notre communauté de jeunes chercheurs!
 
Merci aux organisateurs des JEP et à l'AFCP pour leur soutien.
Back  Top

4-4Le projet European Language Equality

Le projet European Language Equality vise à établir un agenda stratégique concernant la
recherche et l’innovation pour atteindre l’égalité des langues à l’ère numérique en
Europe en 2030. Dans le cadre de ce projet, les partenaires ont produit des rapports
documentant l'état des technologies et ressources pour chaque langue officielle, ainsi
que pour certaines langues non-officielles (D1.4-D1.36). Des états de l'art couvrant
quatre grands domaines ont également été produits (D2.12-D2.16):
- la traduction automatique:
- les technologies vocales
- les technologies pour l'analyse et la compréhension des langues
- les ressources et les bases de connaissance.

Tous ces rapports sont accessibles depuis le site du projet:
https://european-language-equality.eu/deliverables/

Le rapport consacré à l'état des technologies pour la langue française et pour la langue
des signes française a fait l'objet d'une traduction en français. Il est disponible ici:
https://hal.archives-ouvertes.fr/hal-03637784

Back  Top

4-5Distribution Agreement between ELDA and Lexicala for Multilingual Lexical Data Dissemination

Press Release – immediate
Paris, France and Tel Aviv, Israel, October 12, 2023

 

Distribution Agreement between ELDA and Lexicala for Multilingual Lexical Data Dissemination

ELDA and Lexicala by K Dictionaries are delighted to announce their new cooperation on distributing Language Resources for 50 languages.

ELDA is now making available Lexicala’s high-quality lexical data designed to enhance language learning, and support Machine Translation and diverse Natural Language Processing and Artificial Intelligence applications.

The Lexicala resources consist of different groups of datasets. Full descriptions can be found in the ELRA Catalogue of Language Resources under the following links:

  1. GLOBAL Multilingual Lexical Data: a network of lexicographic cores for major world languages, comprising monolingual cores, bilingual pairs, and multilingual combinations for 25 languages.
  2. MULTIGLOSS Multilingual Glossaries: a series of innovative word-to-sense glossaries for over 30 languages into 45 more languages.
  3. Morphological lexicons: extensive morphological lists linking inflected forms to main lemmas for 15 languages.
  4. Parallel Corpora & Domains: parallel corpora for nearly 400 language pairs and numerous multilingual combinations, featuring general language and vertical domain vocabularies.
  5. Biographical & Geographical Names:
  • English BIO Biographical Names: 4,200 dictionary entries regarding prominent persons worldwide.
  • English GEO Geographical Names: 7,200 dictionary entries regarding major locations worldwide.
  • GEOLINGUAL Tables: multilingual tables of over 200 countries and geographical names – including their adjectives, persons, and main languages – in 16 languages.
  • Audio Pronunciation & Phonetic Transcription: human voice recordings of single-word lemmas and multiword expressions, as well as IPA and alternative scripts for 21 languages.

 

 For more information, please write to contact@elda.org.

 About Lexicala

Lexicala by K Dictionaries offers multi-layer lexical data for and across 50 languages, relying on 30-year experience in pedagogical and multilingual lexicography worldwide. Lexicala converges manual content creation and curation with automated data processes and helps to enhance machine translation and other natural language processing applications, as well as language learning and model training.

To find out more about Lexicala, please visit: https://lexicala.com/

 

About ELDA

The Evaluation and Language resources Distribution Agency (ELDA) identifies, collects, markets, and distributes language resources, along with the dissemination of general information in the field of Human Language Technologies (HLT). ELDA has considerable knowledge and skills in HLT applications. ELDA is part of major French, European and international projects in the field of HLT.

To find out more about ELDA, please visit: http://www.elda.org/

Back  Top

4-6Call for membership: ManyLanguages
We are excited to launch ManyLanguages, a globally distributed network of laboratories that helps coordinate Big Team Science studies on human language. 

Our mission is to facilitate the connection between language science researchers to diversify the languages, participants, researchers, and projects represented in the language sciences. We will facilitate the collection of evidence across the language sciences by supporting a distributed laboratory network that is ongoing, diverse, and inclusive. We embrace open science principles by sharing collected data, materials, translations, and other research outputs from the network. We strive to engage research across a broad spectrum of language sciences creating an inclusive and diverse environment for ideas, investigation, and participation.

Join us as a member and learn more about our plans to help advance the language sciences. Currently, we are accepting proposals for big team science projects that replicate experimental linguistic phenomena across many languages. Selected projects will be supported by our team and external experts throughout the entire project.

Find more information here: https://many-languages.com 
Join us as a member here: https://many-languages.com/join.html 

Get in touch with us: many-languages@googlegroups.com 

And follow us on social media for updates: 

Back  Top

4-7The Speech Prosody Conference program 2024

 

  1. The Speech Prosody Conference Program is now available, at https://www.universiteitleiden.nl/sp2024/program .

 

  1. The online lecture series resumes next month, with a talk on speech synthesis needs by Zofia Malisz; details below, also at https://sprosig.org/index.html .  After that the tentative schedule is
  • Gabriel Skantze, KTH, May 15.
  • Simon Roessig, York, September.
  • Sam Tilsen, Cornell, October.
  • Sasha Calhoun, Victoria University of Wellington, November.
  • Robert Xu, Stanford, December.

 

 

The speech synthesis phoneticians need is both realistic and controllable: A survey and a roadmap towards modern synthesis tools for phonetics.
Zofia Malisz, KTH Royal Institute of Technology.
April 17th, 2 pm (Brasilia time). viewing link

ABSTRACT
In the last decade, data and machine learning-driven methods to speech synthesis have greatly improved its quality. So much so, that the realism achievable by current neural synthesisers can rival natural speech. However, modern neural synthesis methods have not yet transferred as tools for experimentation in the speech and language sciences. This is because modern systems still lack the ability to manipulate low-level acoustic characteristics of the signal such as e.g.: formant frequencies.
In this talk, I survey recent advances in speech synthesis and discuss their potential as experimental tools for phonetic research. I argue that speech scientists and speech engineers would benefit from working more with each other again: in particular, in the pursuit of prosodic and acoustic parameter control in neural speech synthesis. I showcase several approaches to fine synthesis control that I have implemented with colleagues: the WavebenderGAN and a system that mimicks the source-filter model of speech production. These systems allow to manipulate formant frequencies and other acoustic parameters with the same (or better) accuracy as e.g.: Praat but with a far superior signal quality.
Finally, I discuss ways to improve synthesis evaluation paradigms, so that not only industry but also speech science experimentation benchmarks are met. My hope is to inspire more students and researchers to take up these research challenges and explore the potential of working at the intersection of the speech technology and speech science.

Outline: 1. I discuss briefly the history of advancements in speech synthesis starting in the formant synthesis era and explain where the improvements came from. 2. I show experiments that I have done that prove modern synthesis is processed not differently than natural speech by humans in a lexical decision task as evidence that the realism (“naturalness”) goal has been largely achieved. 3. I explain how realism came at the expense of controllability. I show how controllability is an indispensable feature for speech synthesis to be adopted in phonetic experimentation. I survey the current state of research on controllability in speech engineering - concentrating on prosodic and formant control. 4. I propose how we can fix this by explaining the work I have done with colleagues on several systems that feature both realism and control. 5. I sketch a roadmap to improve synthesis tools for phonetics - by placing focus on benchmarking systems according to scientific criteria.

  • TBD. Gabriel Skantze, KTH, May 15.
  • TBD. Simon Roessig, York, September.
  • TBD. Sam Tilsen. October.
  • TBD. Sasha Calhoun, November.
  • TBD, Robert Xu, December.

 

Nigel Ward, SProSIG Chair, Professor of Computer Science, University of Texas at El Paso

nigel@utep.edu    https://www.cs.utep.edu/nigel/   

Back  Top

4-8CfProposals for the organization of the 2029 ICASSP

The IEEE Signal Processing Society (SPS) welcomes proposals for the organization of the 2029 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) and invites SPS members to submit a proposal.

 

ICASSP is the world’s largest and most comprehensive technical conference focused on signal processing applications. The series is sponsored by the IEEE Signal Processing Society and has been held annually since 1976. The conference features world-class speakers, tutorials, exhibits, and over 170 lecture and poster sessions. 

       To submit proposals:
https://signalprocessingsociety.org/sites/default/files/uploads/conferences_events/docs/ICASSP_Proposal_Guidelines.pdf?utm_campaign=ICA

To submit a proposal, first familiarize yourself with the SPS Conference Organizer Guidelines. Please read, review, discuss, and build your proposal and plans based around these guidelines. If you’d like more information about ICASSP to help strengthen your proposal development, you can request support by completing the SPS Conference Proposal Information Request Form. Then, please complete and submit the following material by 20 November 2024:


1.       SPS Conference Proposal Pre-Screening Form
2.       SPS Conference Proposal Pre-Screening Budget Summary
3.       Organizing Committee List

 

 

The SPS Conferences Board Executive Subcommittee will assess proposal Pre-screening Forms to determine the final proposing teams.

 

Finalists will be invited to prepare full proposals, host a site visit by SPS representatives in March 2025, and present the proposal at the SPS Conferences Board meeting held at ICASSP 2025 from 06-11 April in Hyderabad, India.

 
 
                       https://app.smartsheet.com/b/form/9a027f4b16564c5a91576c330eee2104?utm_campaign=ICASSP%202029&utm_medium=email

For additional inquiries regarding ICASSP 2029 proposal submission, please email the Signal Processing Society at sps-conf-proposals@ieee.org.

Back  Top

4-9Cf Task proposals at CHiME-9 Challenge
Call for task proposals!
Come and share your ideas about distant microphone speech processing with us!
Task proposals deadline October 25.
 
CHiME is a challenge series whose purpose is to produce scientific advances in the field of distant microphone speech processing. Since CHiME-7, each edition of the CHiME Challenge features independent tasks organised by teams who work within a schedule defined by the CHiME Steering Group. Teams are responsible for managing the tasks, i.e., providing data and instructions, scoring submissions and results, etc. Once the proposal is conditionally accepted, the CHiME Steering Group works to support teams in the development of their task by helping with the design, schedule, webpage/platform and organise a CHiME workshop where the tasks will be presented via regular consultations.
 
For more information:
Back  Top

4-10CfProposals of Conferences and Workshops sponsored by Signal Processing Society (SPS)

The Signal Processing Society (SPS) has implemented an Annual Call for Proposals for many of the technical conferences and workshops solely sponsored by SPS. Proposals are welcome for the organization of the below indicated workshops with final selection occurring in April 2025. 


This Call for Proposals is distributed on behalf of the Signal Processing Society Technical Directions Board in solicitation of proposals for the following workshops: 

  • IEEE Sensor Array and Multichannel Signal Processing Workshop - SAM 2026
  • IEEE Workshop on Signal Processing Systems - SiPS 2026
  • IEEE Spoken Language Technology Workshop - SLT 2026
  • IEEE International Workshop on Information Forensics and Security - WIFS 2026
  • IEEE International Workshop on Machine Learning for Signal Processing - MLSP 2027
  • IEEE International Workshop on Multimedia Signal Processing - MMSP 2027
  • IEEE Statistical Signal Processing Workshop - SSP 2027 

Interested parties will be asked to complete a short Intent to Bid Form by 27 November 2024. Finalists will be selected by 6 December 2024 and will be invited to prepare a full proposal, due 25 March 2025. Visit the Call for Proposals page for more details about each call.


SPS continues to be dedicated to providing opportunities to network with peers from around the world, and to exchange high-quality, rigorously peer-reviewed scientific and technological knowledge. This change is implemented to increase the transparency of the bidding process for our valued Organizers.

 

For additional inquiries regarding proposal submission, please visit the Signal Processing Society website at https://signalprocessingsociety.org/events/conference-call-proposals or email the Signal Processing Society Conferences team at sp.conferences.info@ieee.org.

Back  Top

4-11Bids for Interspeech 2028

Bids for Interspeech 2028


ISCA now invites bids for hosting Interspeech 2028. Interspeech conferences include papers on all the scientific and technological aspects of Speech. More than 1,500 participants from all over the world attend the conference annually to present their work in oral and poster sessions. Several satellite workshops and a Scientific and Industrial Exhibition highly enrich the conference content. Bids must be submitted by members of the scientific community; bids coming directly from PCOs will not be considered. However, we strongly recommend that organizers collaborate with a PCO (including during the preparation of the bid, if you so desire).


Interspeech conferences may be held in any country, although they generally should not occur on the same continent in two consecutive years. After this year’s Interspeech conference in Kos, the next conferences will be held in Rotterdam (The Netherlands) in 2025, Sydney (Australia) in 2026, and São Paolo (Brazil) in 2027.


In order to prepare the bid, please contact the ISCA conference coordinators at conferences@isca-speech.org well in advance of the deadline in order to prepare a high-quality bid. Each bid needs to include:


    - the bidding and budget template

    - a detailed description of the bid

    - other material which might be necessary for evaluating the bid


Guidelines on how to prepare an Interspeech conference can be found here.


Deadline for bid submission: November 1, 2024

Back  Top



 Organisation  Events   Membership   Help 
 > Board  > Interspeech  > Join - renew  > Sitemap
 > Legal documents  > Workshops  > Membership directory  > Contact
 > Logos      > FAQ
       > Privacy policy

© Copyright 2024 - ISCA International Speech Communication Association - All right reserved.

Powered by ISCA