acoustic representation
Recently Published Documents


TOTAL DOCUMENTS

17
(FIVE YEARS 1)

H-INDEX

2
(FIVE YEARS 0)

Electronics ◽  
2020 ◽  
Vol 9 (10) ◽  
pp. 1725
Author(s):  
Gintautas Tamulevičius ◽  
Gražina Korvel ◽  
Anil Bora Yayak ◽  
Povilas Treigys ◽  
Jolita Bernatavičienė ◽  
...  

In this research, a study of cross-linguistic speech emotion recognition is performed. For this purpose, emotional data of different languages (English, Lithuanian, German, Spanish, Serbian, and Polish) are collected, resulting in a cross-linguistic speech emotion dataset with the size of more than 10.000 emotional utterances. Despite the bi-modal character of the databases gathered, our focus is on the acoustic representation only. The assumption is that the speech audio signal carries sufficient emotional information to detect and retrieve it. Several two-dimensional acoustic feature spaces, such as cochleagrams, spectrograms, mel-cepstrograms, and fractal dimension-based space, are employed as the representations of speech emotional features. A convolutional neural network (CNN) is used as a classifier. The results show the superiority of cochleagrams over other feature spaces utilized. In the CNN-based speaker-independent cross-linguistic speech emotion recognition (SER) experiment, the accuracy of over 90% is achieved, which is close to the monolingual case of SER.


Electronics ◽  
2020 ◽  
Vol 9 (7) ◽  
pp. 1125
Author(s):  
Haitao Lang ◽  
Jie Yang

Recently, supervised learning methods have shown promising performance, especially deep neural network-based (DNN) methods, in the application of single-channel speech enhancement. Generally, those approaches extract the acoustic features directly from the noisy speech to train a magnitude-aware target. In this paper, we propose to extract the acoustic features not only from the noisy speech but also from the pre-estimated speech, noise and phase separately, then fuse them into a new complementary feature for the purpose of obtaining more discriminative acoustic representation. In addition, on the basis of learning a magnitude-aware target, we also utilize the fusion feature to learn a phase-aware target, thereby further improving the accuracy of the recovered speech. We conduct extensive experiments, including performance comparison with some typical existing methods, generalization ability evaluation on unseen noise, ablation study, and subjective test by human listener, to demonstrate the feasibility and effectiveness of the proposed method. Experimental results prove that the proposed method has the ability to improve the quality and intelligibility of the reconstructed speech.


2020 ◽  
Vol 6 (s1) ◽  
Author(s):  
Jane Stuart-Smith

AbstractThis paper considers the relative influence on sociophonetic interpretation of /s/ using “static” and “dynamic” acoustic analysis, where dynamic refers to the use of measures which capture the time-varying nature of segmental acoustics, and static to measures which are taken at a single point, or from an average across the sound (Watson and Harrington 1999, Docherty et al. 2015). Static and dynamic Discrete Cosine Transformation (DCT) analyses of spectral Centre of Gravity (CoG) and spectral Slope measures/trajectories were carried out on gendered productions of /s/ and /ʃ/ for a real- and apparent-time 32 speaker sample of spontaneous Glasgow dialect. Results of static CoG measures, reflecting place of articulation, indicate a reduction of gender differentation over time, such that girls born most recently revert to older vernacular (lower frequency) norms. Adding static spectral Slope, reflecting articulatory constriction, shows a change in gendered differentiation, whereby boys born most recently show a gestural shift. The DCT analysis both confirms the static results and also reveals that dynamic characteristics of both sibilants carry key additional prosodic, linguistic and social information for this community. Our results reflect for the first time the usefulness of changing analytical perspectives on /s/, both in terms of acoustic representation (static and dynamic), but also by considering /s/ alongside its partner sibilant /ʃ/ within the context of shifts in the construction of social gender over time.


2018 ◽  
Vol 33 (80) ◽  
pp. 125-144
Author(s):  
Christian Dahl

Christian Dahl: “Battle scenes in the Elizabethan theater”This article analyses the widespread use of staged battle in Elizabethan theater by use of data extracted from Folger Library’s Digital Anthology of Early English Drama. Between 1576 and 1616, hundreds of battle scenes were produced on English stages but although a substantial number is still available for study, only few scholars have recognized their significance. The many battle scenes both attest to the Elizabethans’ vivid interest in history and to the cultural impact of England’s increasing military engagement on the Continent and in Ireland at the end of Elizabeth’s reign. It is often assumed that histories and battle scenes were particularly popular in the 1590’ies and then fell out of fashion early in the 17th century, but the article demonstrates that staged war remained a frequent occurrence in the first two decades of the century and never disappeared entirely. The article discusses visual and, in particular, acoustic representation of warfare based on the evidence of surviving plays and other documents. The article will also (very) briefly sketch the narrative development of battle scenes that took place in the 1590ies.


eLife ◽  
2018 ◽  
Vol 7 ◽  
Author(s):  
Manfred Kössl ◽  
Julio Hechavarría

The acoustic representation of the outside world in the midbrain of a bat becomes more precise as it uses double clicks to locate closer objects.


2018 ◽  
Vol 231 ◽  
pp. 05002 ◽  
Author(s):  
Karolina Marciniuk ◽  
Maciej Blaszke ◽  
Bożena Kostek

The subject of this research is showing the performance of an automatic acoustic road monitoring system proposed by the authors. The main goal of the study is describing road traffic by means of an acoustic representation and testing effectiveness of traffic flow sensors. Evaluation metrics of the road conditions such as velocity of the traffic flow, its structure and weather condition are presented along with acoustic descriptors derived from the audio signal analysis. Accuracy of emergency vehicles pass by detection based on acoustic monitoring is also briefly described.


Sign in / Sign up

Export Citation Format

Share Document