- Part of the Prosody, Phonology and Phonetics book series (PRPHPH)
ISCApad #270 |
Friday, December 11, 2020 by Chris Wellekens |
5-1-1 | Emmanuel Vincent (Editor), Tuomas Virtanen (Editor), Sharon Gannot (Editor), 'Audio Source Separation and Speech Enhancement', Wiley Emmanuel Vincent (Editor), Tuomas Virtanen (Editor), Sharon Gannot (Editor), Audio Source Separation and Speech Enhancement:
ISBN: 978-1-119-27989-1 October 2018 504 pages
| |||||||
5-1-2 | Jen-Tzung Chien, 'Source Separation and Machine Learning', Academic Press Jen-Tzung Chien, 'Source Separation and Machine Learning', Academic Press
| |||||||
5-1-3 | Ingo Feldhausen, « Methods in prosody: A Romance language perspective », Language Science Press (open access) Nous sommes heureux de vous annoncer la parution d'un recueil validé par un comité de lecture et consacré aux méthodes de recherche en prosodie. Cet ouvrage est intitulé « Methods in prosody: A Romance language perspective ». Il est publié par Language Science Press, une maison d’édition open access. Le livre peut-être téléchargé gratuitement en cliquant sur le lien suivant : http://langsci-press.org/catalog/book/183 La table des matières est la suivante : --------------------------------------------------------------------------------------------------------- Introduction Foreword I Large corpora and spontaneous speech 1) Using large corpora and computational tools to describe prosody: An 2) Intonation of pronominal subjects in Porteño Spanish: Analysis of II Approaches to prosodic analysis 3) Multimodal analyses of audio-visual information: Some methods and 4) The realizational coefficient: Devising a method for empirically 5) On the role of prosody in disambiguating wh-exclamatives and III Elicitation methods 6) The Discourse Completion Task in Romance prosody research: Status 7) Describing the intonation of speech acts in Brazilian Portuguese: Indexes 263 --------------------------------------------------------------------------------------------------------- N'hésitez pas à diffuser la parution de cet ouvrage auprès de vos collègues qui pourraient s'y intéresser. Bien cordialement, Ingo Feldhausen
| |||||||
5-1-4 | Nigel Ward, 'Prosodic Patterns in English Conversation', Cambridge University Press, 2019 Prosodic Patterns in English Conversation Nigel G. Ward, Professor of Computer Science, University of Texas at El Paso Cambridge University Press, 2019.
Spoken language is more than words: it includes the prosodic features and patterns that speakers use, subconsciously, to frame meanings and achieve interactional goals. Thanks to the application of simple processing techniques to spoken dialog corpora, this book goes beyond intonation to describe how pitch, timing, intensity and voicing properties combine to form meaningful temporal configurations: prosodic constructions. Combining new findings with hitherto-scattered observations from diverse research traditions, this book enumerates twenty of the principal prosodic constructions of English.
http://www.cambridge.org/ward/ nigel@utep.edu http://www.cs.utep.edu/nigel/
| |||||||
5-1-5 | J.H.Esling, Scott R.Moisik, Allison Benner, Lise Crevier-Buchman, 'Voice Quality: the Laryngeal Articulator Model', Cambridge University Press Voice Quality The Laryngeal Articulator Model Hardback 978-1-108-49842-5 John H. Esling, University of Victoria, British Columbia Scott R. Moisik, Nanyang Technological University, Singapore Allison Benner, University of Victoria, British Columbia Lise Crevier-Buchman, Centre National de la Recherche Scientifique (CNRS), Paris The first description of voice quality production in forty years, this book provides a new framework for its study: The Laryngeal Articulator Model. Informed by instrumental examinations of the laryngeal articulatory mechanism, it revises our understanding of articulatory postures to explain the actions, vibrations and resonances generated in the epilarynx and pharynx. It focuses on the long-term auditory-articulatory component of accent in the languages of the world, explaining how voice quality relates to segmental and syllabic sounds. Phonetic illustrations of phonation types and of laryngeal and oral vocal tract articulatory postures are provided. Extensive video and audio material is available on a companion website. The book presents computational simulations, the laryngeal and voice quality foundations of infant speech acquisition, speech/voice disorders and surgeries that entail compensatory laryngeal articulator adjustment, and an exploration of the role of voice quality in sound change and of the larynx in the evolution of speech.
1. Voice and voice quality; 2. Voice quality classification; 3. Instrumental case studies and computational simulations of voice quality; 4. Linguistic, paralinguistic and extralinguistic illustrations of voice quality; 5. Phonological implications of voice quality theory; 6. Infant acquisition of speech and voice quality; 7. Clinical illustrations of voice quality; 8. Laryngeal articulation and voice quality in sound change, language ontogeny.
| |||||||
5-1-6 | Albert di Cristo,' Les langues naturelles'. HAL archive ouverte Albert di Cristo, les langues naturelles. https://hal-amu.archives-ouvertes.fr/hal-02149640 Cet ouvrage constitue la première partie d?un vaste travail dédié à l?étude des façons dont les langues naturelles conditionnent l?information et au rôle que joue la prosodie dans l?expression de ces conditionnements. Cette première partie se propose d?analyser, sous ses divers aspects (principalement d'ordre épistémologiques), la notion de structure informationnelle, notamment dans ses relations avec la grammaire et d?examiner, dans le détail, les déterminants qui forment l?armature de cette structure. Dans cette perspective, les discussions portent, outre sur les notions de thème, de topique et de « given », sur celles de focus, de focalisation et de contraste, qui font l?objet d?analyses approfondies. Les discussions s?attachent à appréhender ces notions, à la fois dans l?optique de leurs propriétés formelles, de leur fonctionnalité et des significations qu?elles contribuent à délivrer. Un chapitre entier de cette première partie est consacré à l?étude du questionnement et à la manière dont l?organisation de l?information est gérée dans l?exercice de cette activité. L?ouvrage est doté d?une bibliographie qui comporte plus de deux mille références. Cet ouvrage sera complété par une 2ème partie, en cours de rédaction, qui traitera essentiellement de la prosodie et de son rôle dans les conditionnements de l'information.
| |||||||
5-1-7 | Benjamin Weiss, 'Talker Quality in Human and Machine Interaction - Modeling the Listener’s Perspective in Passive and Interactive Scenarios'. T-Labs Series in Telecommunication Services. Springer Nature, Cham. (2020) Benjamin Weiss (2020): 'Talker Quality in Human and Machine Interaction - Modeling the Listener’s Perspective in Passive and Interactive Scenarios'. T-Labs Series in Telecommunication Services. Springer Nature, Cham.
https://rd.springer.com/book/10.1007/978-3-030-22769-2
| |||||||
5-1-8 | W.F.Katz, P.F.Assman, 'The Routledge Handbook of Phonetics', Routledge.
| |||||||
5-1-9 | Proceedings of SLTU-CCURL2020 Dear all, we are very happy to announce that the SLTU-CCURL2020 Proceedings are available online: https://lrec2020.lrec-conf.org/media/proceedings/Workshops/Books/SLTUCCURLbook.pdf
This year, LREC2020 would have featured an extraordinary event: the first joint SLTU-CCURL2020 Workshop, which was planned as a two-day workshop, with 54 papers accepted either as oral and poster presentations.
The workshop program was enriched by two tutorials and two keynote speeches.
We will miss the presentations, the discussions and the overall stimulating environment very deeply.
We are thankful to ELRA and ISCA for their support to the workshop, to our Google sponsor and to the 60 experts of the Program Committee, who worked tirelessly in order to help us to select the best papers representing a wide perspective over NLP, speech and computational linguistics addressing less-resource languages.
Looking forward to better times when we will be able to meet in person again, we hope that you will find these workshop proceedings relevant and stimulating for your own research.
With our best wishes,
Claudia Soria, Laurent Besacier, Dorothee Beermann, and Sakriani Sakti
| |||||||
5-1-10 | B.Weiss, J.Trouvain, M. Barkat-Defradas, J.J.Ohala, ' Voice Attractiveness', SpringerVoice AttractivenessStudies on Sexy, Likable, and Charismatic Speakers
|
5-2-1 | Linguistic Data Consortium (LDC) update (November 2020)
In this newsletter: Join LDC for Membership Year 2021
New publications:
Global TIMIT Learner Simple English is distributed via web download. * (2) LORELEI Ukrainian Representative Language Pack consists of Ukrainian monolingual text, Ukrainian-English parallel and comparable text, annotations, supplemental resources, and related software tools developed by LDC for the DARPA LORELEI program.
Approximately 75,000 words were annotated for named entities and up to 50,000 words contain additional annotation, including situation frames (identifying entities, needs and issues) and entity linking and detection. * (3) TAC KBP Event Argument – Comprehensive Training and Evaluation Data 2016-2017 was developed by LDC and contains training and evaluation data produced in support of the 2016 TAC KBP Event Argument Linking Pilot and Evaluation tasks and the 2017 Event Argument Linking Training Evaluation task.
Membership Coordinator University of Pennsylvania T: +1-215-573-1275 E: ldc@ldc.upenn.edu M: 3600 Market St. Suite 810 Philadelphia, PA 19104
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-2 | ELRA - Language Resources Catalogue - Update (November 2020) We are happy to announce that 1 new Speech resource is now available in our catalogue.
ISLRN: 425-664-403-057-4
Ahoslabi was built within the frame of the RESTORE project (?Restauración, almacenamiento y rehabilitación de la voz?) (restrictions apply). The database primarily consists of recordings of 31 laryngectomees (27 males and 4 females) pronouncing 100 phonetically balanced sentences. The total size of the recordings amount 10h48min for 1.16 Gb. Esophageal voices were recorded in a soundproof recording cubicle with a Neuman microphone. Additionally, it includes parallel recordings of the sentences by 9 healthy speakers (6 males and 3 females) to facilitate speech processing tasks that require small parallel corpora, such as voice conversion or synthetic speech adaptation. A pronunciation lexicon in SAMPA is also provided. For more information, see: http://catalog.elra.info/en-us/repository/browse/ELRA-S0413 For more information on the catalogue, please contact Valérie Mapelli mailto:mapelli@elda.org If you would like to enquire about having your resources distributed by ELRA, please do not hesitate to contact us. Visit the Universal Catalogue: http://universal.elra.info Archives of ELRA Language Resources Catalogue Updates: http://www.elra.info/en/catalogues/language-resources-announcements
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-3 | Speechocean – update (August 2019)
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-4 | Google 's Language Model benchmark A LM benchmark is available at:https://github.com/ciprian-chelba/1-billion-word-language-modeling-benchmark
Here is a brief description of the project.
'The purpose of the project is to make available a standard training and test setup for language modeling experiments. The training/held-out data was produced from a download at statmt.org using a combination of Bash shell and Perl scripts distributed here. This also means that your results on this data set are reproducible by the research community at large. Besides the scripts needed to rebuild the training/held-out data, it also makes available log-probability values for each word in each of ten held-out data sets, for each of the following baseline models:
ArXiv paper: http://arxiv.org/abs/1312.3005
Happy benchmarking!'
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-5 | Forensic database of voice recordings of 500+ Australian English speakers Forensic database of voice recordings of 500+ Australian English speakers
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-6 | Audio and Electroglottographic speech recordings
Audio and Electroglottographic speech recordings from several languages We are happy to announce the public availability of speech recordings made as part of the UCLA project 'Production and Perception of Linguistic Voice Quality'. http://www.phonetics.ucla.edu/voiceproject/voice.html Audio and EGG recordings are available for Bo, Gujarati, Hmong, Mandarin, Black Miao, Southern Yi, Santiago Matatlan/ San Juan Guelavia Zapotec; audio recordings (no EGG) are available for English and Mandarin. Recordings of Jalapa Mazatec extracted from the UCLA Phonetic Archive are also posted. All recordings are accompanied by explanatory notes and wordlists, and most are accompanied by Praat textgrids that locate target segments of interest to our project. Analysis software developed as part of the project – VoiceSauce for audio analysis and EggWorks for EGG analysis – and all project publications are also available from this site. All preliminary analyses of the recordings using these tools (i.e. acoustic and EGG parameter values extracted from the recordings) are posted on the site in large data spreadsheets. All of these materials are made freely available under a Creative Commons Attribution-NonCommercial-ShareAlike-3.0 Unported License. This project was funded by NSF grant BCS-0720304 to Pat Keating, Abeer Alwan and Jody Kreiman of UCLA, and Christina Esposito of Macalester College. Pat Keating (UCLA)
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-7 | EEG-face tracking- audio 24 GB data set Kara One, Toronto, Canada We are making 24 GB of a new dataset, called Kara One, freely available. This database combines 3 modalities (EEG, face tracking, and audio) during imagined and articulated speech using phonologically-relevant phonemic and single-word prompts. It is the result of a collaboration between the Toronto Rehabilitation Institute (in the University Health Network) and the Department of Computer Science at the University of Toronto.
In the associated paper (abstract below), we show how to accurately classify imagined phonological categories solely from EEG data. Specifically, we obtain up to 90% accuracy in classifying imagined consonants from imagined vowels and up to 95% accuracy in classifying stimulus from active imagination states using advanced deep-belief networks.
Data from 14 participants are available here: http://www.cs.toronto.edu/~complingweb/data/karaOne/karaOne.html.
If you have any questions, please contact Frank Rudzicz at frank@cs.toronto.edu.
Best regards, Frank
PAPER Shunan Zhao and Frank Rudzicz (2015) Classifying phonological categories in imagined and articulated speech. In Proceedings of ICASSP 2015, Brisbane Australia ABSTRACT This paper presents a new dataset combining 3 modalities (EEG, facial, and audio) during imagined and vocalized phonemic and single-word prompts. We pre-process the EEG data, compute features for all 3 modalities, and perform binary classi?cation of phonological categories using a combination of these modalities. For example, a deep-belief network obtains accuracies over 90% on identifying consonants, which is signi?cantly more accurate than two baseline supportvectormachines. Wealsoclassifybetweenthedifferent states (resting, stimuli, active thinking) of the recording, achievingaccuraciesof95%. Thesedatamaybeusedtolearn multimodal relationships, and to develop silent-speech and brain-computer interfaces.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-8 | TORGO data base free for academic use. In the spirit of the season, I would like to announce the immediate availability of the TORGO database free, in perpetuity for academic use. This database combines acoustics and electromagnetic articulography from 8 individuals with speech disorders and 7 without, and totals over 18 GB. These data can be used for multimodal models (e.g., for acoustic-articulatory inversion), models of pathology, and augmented speech recognition, for example. More information (and the database itself) can be found here: http://www.cs.toronto.edu/~complingweb/data/TORGO/torgo.html.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-9 | Datatang Datatang is a global leading data provider that specialized in data customized solution, focusing in variety speech, image, and text data collection, annotation, crowdsourcing services.
Summary of the new datasets (2018) and a brief plan for 2019.
? Speech data (with annotation) that we finished in 2018
?2019 ongoing speech project
On top of the above, there are more planed speech data collections, such as Japanese speech data, children`s speech data, dialect speech data and so on.
What is more, we will continually provide those data at a competitive price with a maintained high accuracy rate.
If you have any questions or need more details, do not hesitate to contact us jessy@datatang.com
It would be possible to send you with a sample or specification of the data.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-10 | Fearless Steps Corpus (University of Texas, Dallas) Fearless Steps Corpus John H.L. Hansen, Abhijeet Sangwan, Lakshmish Kaushik, Chengzhu Yu Center for Robust Speech Systems (CRSS), Eric Jonsson School of Engineering, The University of Texas at Dallas (UTD), Richardson, Texas, U.S.A.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-11 | SIWIS French Speech Synthesis Database The SIWIS French Speech Synthesis Database includes high quality French speech recordings and associated text files, aimed at building TTS systems, investigate multiple styles, and emphasis. A total of 9750 utterances from various sources such as parliament debates and novels were uttered by a professional French voice talent. A subset of the database contains emphasised words in many different contexts. The database includes more than ten hours of speech data and is freely available.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-12 | JLCorpus - Emotional Speech corpus with primary and secondary emotions JLCorpus - Emotional Speech corpus with primary and secondary emotions:
For further understanding the wide array of emotions embedded in human speech, we are introducing an emotional speech corpus. In contrast to the existing speech corpora, this corpus was constructed by maintaining an equal distribution of 4 long vowels in New Zealand English. This balance is to facilitate emotion related formant and glottal source feature comparison studies. Also, the corpus has 5 secondary emotions along with 5 primary emotions. Secondary emotions are important in Human-Robot Interaction (HRI), where the aim is to model natural conversations among humans and robots. But there are very few existing speech resources to study these emotions,and this work adds a speech corpus containing some secondary emotions. Please use the corpus for emotional speech related studies. When you use it please include the citation as: Jesin James, Li Tian, Catherine Watson, 'An Open Source Emotional Speech Corpus for Human Robot Interaction Applications', in Proc. Interspeech, 2018. To access the whole corpus including the recording supporting files, click the following link: https://www.kaggle.com/tli725/jl-corpus, (if you have already installed the Kaggle API, you can type the following command to download: kaggle datasets download -d tli725/jl-corpus) Or if you simply want the raw audio+txt files, click the following link: https://www.kaggle.com/tli725/jl-corpus/downloads/Raw%20JL%20corpus%20(unchecked%20and%20unannotated).rar/4 The corpus was evaluated by a large scale human perception test with 120 participants. The link to the survey are here- For Primary emorion corpus: https://auckland.au1.qualtrics.com/jfe/form/SV_8ewmOCgOFCHpAj3 For Secondary emotion corpus: https://auckland.au1.qualtrics.com/jfe/form/SV_eVDINp8WkKpsPsh These surveys will give an overall idea about the type of recordings in the corpus. The perceptually verified and annotated JL corpus will be given public access soon.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-13 | OPENGLOT –An open environment for the evaluation of glottal inverse filtering OPENGLOT –An open environment for the evaluation of glottal inverse filtering
OPENGLOT is a publically available database that was designed primarily for the evaluation of glottal inverse filtering algorithms. In addition, the database can be used in evaluating formant estimation methods. OPENGLOT consists of four repositories. Repository I contains synthetic glottal flow waveforms, and speech signals generated by using the Liljencrants–Fant (LF) waveform as an excitation, and an all-pole vocal tract model. Repository II contains glottal flow and speech pressure signals generated using physical modelling of human speech production. Repository III contains pairs of glottal excitation and speech pressure signal generated by exciting 3D printed plastic vocal tract replica with LF excitations via a loudspeaker. Finally, Repository IV contains multichannel recordings (speech pressure signal, EGG, high-speed video of the vocal folds) from natural production of speech.
OPENGLOT is available at: http://research.spa.aalto.fi/projects/openglot/
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-14 | Corpus Rhapsodie Nous sommes heureux de vous annoncer la publication d¹un ouvrage consacré
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-15 | The My Science Tutor Children?s Conversational Speech Corpus (MyST Corpus) , Boulder Learning Inc. The My Science Tutor Children?s Conversational Speech Corpus (MyST Corpus) is the world?s largest English children?s speech corpus. It is freely available to the research community for research use. Companies can acquire the corpus for $10,000. The MyST Corpus was collected over a 10-year period, with support from over $9 million in grants from the US National Science Foundation and Department of Education, awarded to Boulder Learning Inc. (Wayne Ward, Principal Investigator). The MyST corpus contains speech collected from 1,374 third, fourth and fifth grade students. The students engaged in spoken dialogs with a virtual science tutor in 8 areas of science. A total of 11,398 student sessions of 15 to 20 minutes produced a total of 244,069 utterances. 42% of the utterances have been transcribed at the word level. The corpus is partitioned into training and test sets to support comparison of research results across labs. All parents and students signed consent forms, approved by the University of Colorado?s Institutional Review Board, that authorize distribution of the corpus for research and commercial use. The MyST children?s speech corpus contains approximately ten times as many spoken utterances as all other English children?s speech corpora combined (see https://en.wikipedia.org/wiki/List_of_children%27s_speech_corpora). Additional information about the corpus, and instructions for how to acquire the corpus (and samples of the speech data) can be found on the Boulder Learning Web site at http://boulderlearning.com/request-the-myst-corpus/.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-16 | HARVARD speech corpus - native British English speaker
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-17 | Magic Data Technology Kid Voice TTS Corpus in Mandarin Chinese (November 2019) Magic Data Technology Kid Voice TTS Corpus in Mandarin Chinese
Magic Data Technology is one of the leading artificial intelligence data service providers in the world. The company is committed to providing a wild range of customized data services in the fields of speech recognition, intelligent imaging and Natural Language Understanding.
This corpus was recorded by a four-year-old Chinese girl originally born in Beijing China. This time we published 15-minute speech data from the corpus for non-commercial use.
The contents and the corresponding descriptions of the corpus:
The corpus aims to help researchers in the TTS fields. And it is part of a much bigger dataset (2.3 hours MAGICDATA Kid Voice TTS Corpus in Mandarin Chinese) which was recorded in the same environment. This is the first time to publish this voice!
Please note that this corpus has got the speaker and her parents’ authorization.
Samples are available. Do not hesitate to contact us for any questions. Website: http://www.imagicdatatech.com/index.php/home/dataopensource/data_info/id/360 E-mail: business@magicdatatech.com
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-18 | FlauBERT: a French LM Here is FlauBERT: a French LM learnt (with #CNRS J-Zay supercomputer) on a large and heterogeneous corpus. Along with it comes FLUE (evaluation setup for French NLP). FlauBERT was successfully applied to complex tasks (NLI, WSD, Parsing). More on https://github.com/getalp/Flaubert
More details on this online paper: https://arxiv.org/abs/1912.05372
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-19 | ELRA-S0408 SpeechTera Pronunciation Dictionary ELRA-S0408 Speechtera Pronunciation Dictionary ISLRN: 645-563-102-594-8
The SpeechTera Pronunciation Dictionary is a machine-readable pronunciation dictionary for Brazilian Portuguese and comprises 737,347 entries. Its phonetic transcription is based on 13 linguistics varieties spoken in Brazil and contains the pronunciation of the frequent word forms found in the transcription data of the SpeechTera's speech and text database (literary, newspaper, movies, miscellaneous). Each one of the thirteen dialects comprises 56,719 entries. For more information, see: http://catalog.elra.info/en-us/repository/browse/ELRA-S0408/ For more information on the catalogue, please contact Valérie Mapelli mailto:mapelli@elda.org If you would like to enquire about having your resources distributed by ELRA, please do not hesitate to contact us. Visit our On-line Catalogue: http://catalog.elra.info
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-20 | Ressources of ELRC Network Paris, France, April 23, 2020 ELRA is happy to announce that Language Resources collected within the ELRC Network, funded by the European Commission, are now available from the ELRA Catalogue of Language Resources. For more information on the catalogue, please contact Valérie Mapelli
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-21 | Language Resources distribution agreement between ELRA and SpeechOcean Press Release - Immediate ELRA and SpeechOcean signed a new Language Resources distribution agreement. On behalf of ELRA, ELDA acts as the distribution agency for SpeechOcean since 2007 and incorporated 46 new speech resources to the ELRA Catalogue of Language Resources catalogue. Those resources were designed and collected to boost Speech Recognition. They cover the following languages:
To find out more about SpeechOcean, please visit the website: http://www.speechocean.com To find out more about ELRA, please visit the website: http://www.elra.info
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-22 | Sharing Language Ressourses via ELRA ELRA recognises the importance of sharing Language Resources (LRs) and making them available to the community. Since the 2014 edition of LREC, the Language Resources and Evaluation Conference, participants have been offered the possibility to share their LRs (data, tools, web-services, etc.) when submitting a paper, uploading them in a special LREC repository set up by ELRA. This effort of sharing LRs, linked to the LRE Map initiative (https://lremap.elra.info) for their description, contributes to creating a common repository where everyone can deposit and share data. The LREC initiative 'Share your LRs' was launched in 2014 in Reykjavik. It was successfully continued in 2016 in Portoro? and 2018 in Miyazaki. Corresponding repositories are available here:
For more information and/or questions, please write to contact@elda.org.
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
5-2-23 | ELRA announces that MEDIA data are now available for free for academic research ELRA announces that MEDIA data are now available for free for academic research Further to the request of the HLT French community to foster evaluation activities for man-machine dialogue systems for French language, ELRA has decided to provide a free access to the MEDIA speech corpora and evaluation package for academic research purposes. The MEDIA data can be found in the ELRA Catalogue under the following references: Data available from the ELRA Catalogue can be obtained easily by contacting ELRA. The MEDIA project was carried out within the framework of Technolangue, the French national research programme funded by the French Ministry of Research and New Technologies (MRNT) with the objective of running a campaign for the evaluation of man-machine dialogue systems for French. The campaign was distributed over two actions: an evaluation taking into account the dialogue context and an evaluation not taking into account the dialogue context. PortMedia was a follow up project supported by the French Research Agency (ANR). The French and Italian corpus was produced by ELDA, with the same paradigm and specifications as the MEDIA speech database but on a different domain. For more information and/or questions, please write to contact@elda.org. *** About ELRA *** To find out more about ELRA and its respective catalogue, please visit: http://www.elra.info and http://catalogue.elra.info
|
5-3-1 | Cantor Digitalis, an open-source real-time singing synthesizer controlled by hand gestures. We are glad to announce the public realease of the Cantor Digitalis, an open-source real-time singing synthesizer controlled by hand gestures. It can be used e.g. for making music or for singing voice pedagogy. A wide variety of voices are available, from the classic vocal quartet (soprano, alto, tenor, bass), to the extreme colors of childish, breathy, roaring, etc. voices. All the features of vocal sounds are entirely under control, as the synthesis method is based on a mathematic model of voice production, without prerecording segments. The instrument is controlled using chironomy, i.e. hand gestures, with the help of interfaces like stylus or fingers on a graphic tablet, or computer mouse. Vocal dimensions such as the melody, vocal effort, vowel, voice tension, vocal tract size, breathiness etc. can easily and continuously be controlled during performance, and special voices can be prepared in advance or using presets. Check out the capabilities of Cantor Digitalis, through performances extracts from the ensemble Chorus Digitalis: http://youtu.be/_LTjM3Lihis?t=13s. In pratice, this release provides:
Regards,
The Cantor Digitalis team (who loves feedback — cantordigitalis@limsi.fr) Christophe d'Alessandro, Lionel Feugère, Olivier Perrotin http://cantordigitalis.limsi.fr/
| |||||
5-3-2 | MultiVec: a Multilingual and MultiLevel Representation Learning Toolkit for NLP
We are happy to announce the release of our new toolkit “MultiVec” for computing continuous representations for text at different granularity levels (word-level or sequences of words). MultiVec includes Mikolov et al. [2013b]’s word2vec features, Le and Mikolov [2014]’s paragraph vector (batch and online) and Luong et al. [2015]’s model for bilingual distributed representations. MultiVec also includes different distance measures between words and sequences of words. The toolkit is written in C++ and is aimed at being fast (in the same order of magnitude as word2vec), easy to use, and easy to extend. It has been evaluated on several NLP tasks: the analogical reasoning task, sentiment analysis, and crosslingual document classification. The toolkit also includes C++ and Python libraries, that you can use to query bilingual and monolingual models.
The project is fully open to future contributions. The code is provided on the project webpage (https://github.com/eske/multivec) with installation instructions and command-line usage examples.
When you use this toolkit, please cite:
@InProceedings{MultiVecLREC2016, Title = {{MultiVec: a Multilingual and MultiLevel Representation Learning Toolkit for NLP}}, Author = {Alexandre Bérard and Christophe Servan and Olivier Pietquin and Laurent Besacier}, Booktitle = {The 10th edition of the Language Resources and Evaluation Conference (LREC 2016)}, Year = {2016}, Month = {May} }
The paper is available here: https://github.com/eske/multivec/raw/master/docs/Berard_and_al-MultiVec_a_Multilingual_and_Multilevel_Representation_Learning_Toolkit_for_NLP-LREC2016.pdf
Best regards,
Alexandre Bérard, Christophe Servan, Olivier Pietquin and Laurent Besacier
| |||||
5-3-3 | An android application for speech data collection LIG_AIKUMA We are pleased to announce the release of LIG_AIKUMA, an android application for speech data collection, specially dedicated to language documentation. LIG_AIKUMA is an improved version of the Android application (AIKUMA) initially developed by Steven Bird and colleagues. Features were added to the app in order to facilitate the collection of parallel speech data in line with the requirements of a French-German project (ANR/DFG BULB - Breaking the Unwritten Language Barrier).
The resulting app, called LIG-AIKUMA, runs on various mobile phones and tablets and proposes a range of different speech collection modes (recording, respeaking, translation and elicitation). It was used for field data collections in Congo-Brazzaville resulting in a total of over 80 hours of speech.
Users who just want to use the app without access to the code can download it directly from the forge direct link: https://forge.imag.fr/frs/download.php/706/MainActivity.apk
Code is also available on demand (contact elodie.gauthier@imag.fr and laurent.besacier@imag.fr).
More details on LIG_AIKUMA can be found on the following paper: http://www.sciencedirect.com/science/article/pii/S1877050916300448
| |||||
5-3-4 | Web services via ALL GO from IRISA-CNRS It is our pleasure to introduce A||GO (https://allgo.inria.fr/ or http://allgo.irisa.fr/), a platform providing a collection of web-services for the automatic analysis of various data, including multimedia content across modalities. The platform builds on the back-end web service deployment infrastructure developed and maintained by Inria?s Service for Experimentation and Development (SED). Originally dedicated to multimedia content, A||GO progressively broadened to other fields such as computational biology, networks and telecommunications, computational graphics or computational physics.
| |||||
5-3-5 | Clickable map - Illustrations of the IPA Clickable map - Illustrations of the IPA
| |||||
5-3-6 | LIG-Aikuma running on mobile phones and tablets
| |||||
5-3-7 | Python Library Nous sommes heureux d'annoncer la mise à disposition du public de la
première bibliothèque en langage Python pour convertir des nombres écrits en
français en leur représentation en chiffres.
L'analyseur est robuste et est capable de segmenter et substituer les expressions
de nombre dans un flux de mots, comme une conversation par exemple. Il reconnaît les différentes
variantes de la langue (quantre-vingt-dix / nonante?) et traduit aussi bien les
ordinaux que les entiers, les nombres décimaux et les séquences formelles (n° de téléphone, CB?).
Nous espérons que cet outil sera utile à celles et ceux qui, comme nous, font du traitment
du langage naturel en français.
Cette bibliothèque est diffusée sous license MIT qui permet une utilisation très libre.
Sources : https://github.com/allo-media/text2num
|