Mongolian speech corpus for text-to-speech development

Author(s):  
Chatchawarn Hansakunbuntheung ◽  
Ausdang Thangthai ◽  
Nattanun Thatphithakkul ◽  
Altangerel Chagnaa
Ifost ◽  
2013 ◽  
Author(s):  
Chagnaa Altangerel ◽  
Kerey Esenbek ◽  
Jaimai Purev

Author(s):  
Marc Freixes ◽  
Francesc Alías ◽  
Joan Claudi Socoró

AbstractText-to-speech (TTS) synthesis systems have been widely used in general-purpose applications based on the generation of speech. Nonetheless, there are some domains, such as storytelling or voice output aid devices, which may also require singing. To enable a corpus-based TTS system to sing, a supplementary singing database should be recorded. This solution, however, might be too costly for eventual singing needs, or even unfeasible if the original speaker is unavailable or unable to sing properly. This work introduces a unit selection-based text-to-speech-and-singing (US-TTS&S) synthesis framework, which integrates speech-to-singing (STS) conversion to enable the generation of both speech and singing from an input text and a score, respectively, using the same neutral speech corpus. The viability of the proposal is evaluated considering three vocal ranges and two tempos on a proof-of-concept implementation using a 2.6-h Spanish neutral speech corpus. The experiments show that challenging STS transformation factors are required to sing beyond the corpus vocal range and/or with notes longer than 150 ms. While score-driven US configurations allow the reduction of pitch-scale factors, time-scale factors are not reduced due to the short length of the spoken vowels. Moreover, in the MUSHRA test, text-driven and score-driven US configurations obtain similar naturalness rates of around 40 for all the analysed scenarios. Although these naturalness scores are far from those of vocaloid, the singing scores of around 60 which were obtained validate that the framework could reasonably address eventual singing needs.


2019 ◽  
Vol 53 (3) ◽  
pp. 419-447
Author(s):  
Humberto M. Torres ◽  
Jorge A. Gurlekian ◽  
Diego A. Evin ◽  
Christian G. Cossio Mercado

2012 ◽  
Vol 21 (2) ◽  
pp. 60-71 ◽  
Author(s):  
Ashley Alliano ◽  
Kimberly Herriger ◽  
Anthony D. Koutsoftas ◽  
Theresa E. Bartolotta

Abstract Using the iPad tablet for Augmentative and Alternative Communication (AAC) purposes can facilitate many communicative needs, is cost-effective, and is socially acceptable. Many individuals with communication difficulties can use iPad applications (apps) to augment communication, provide an alternative form of communication, or target receptive and expressive language goals. In this paper, we will review a collection of iPad apps that can be used to address a variety of receptive and expressive communication needs. Based on recommendations from Gosnell, Costello, and Shane (2011), we describe the features of 21 apps that can serve as a reference guide for speech-language pathologists. We systematically identified 21 apps that use symbols only, symbols and text-to-speech, and text-to-speech only. We provide descriptions of the purpose of each app, along with the following feature descriptions: speech settings, representation, display, feedback features, rate enhancement, access, motor competencies, and cost. In this review, we describe these apps and how individuals with complex communication needs can use them for a variety of communication purposes and to target a variety of treatment goals. We present information in a user-friendly table format that clinicians can use as a reference guide.


Sign in / Sign up

Export Citation Format

Share Document