sound file
Recently Published Documents


TOTAL DOCUMENTS

18
(FIVE YEARS 0)

H-INDEX

2
(FIVE YEARS 0)

2020 ◽  
Vol 8 (3) ◽  
pp. 228-233
Author(s):  
Gilbert E. Bueno ◽  
Kristine A. Valenzuela ◽  
Edwin R. Arboleda

Cacao pod's ideal harvesting time is when it is about to be ripe. Immature harvest would result in hard cacao beans not suitable for fermentation, while overripe cacao pods lead to fungal-infected, defective, and poor-quality yields. The demand for high-quality cacao products is expected to rise due to advancing technology in the present. Pre-harvesting needs to provide optimal identification of which amongst the pods are ripened enough and ready for the next stage of the cacao process. This paper recommends a technique to determine the ripeness of cacao. Nine hundred thirty-three cacao samples were used to collect thumping audio data at five different pod's exocarp locations. Each sound file is 1 second long, creating 4665 cacao sound file datasets at 16kHz sample rate and 16-bit audio bit depth. The process of the Mel-Frequency Cepstral Coefficient Spectogram was then applied to extract recognizable features for the training process. The deep learning method integrated was a convolutional neural network (CNN) to classify the cacao sound successfully. The experimental design model's output exhibits an accuracy of 97.50 % for the training data and 97.13 % for the validation data. While the overall accuracy mean of the classification system is 97.46 %, whether the cacao is unripe or ripe.


In a country like India, wide variety of fruits are available. Fruits plays an important role in the health of human beings and naturally health improves, if the quality of the fruit is good. Grading of the watermelon quality helps the consumers and vendors. The proposed work is to classify the watermelons based on the sound. Sound file dataset is created manually by tapping the watermelon and recording the sound. Dataset consist of different types of watermelon. For this, different size, colour and shape of the watermelons are used. Features are extracted from the sound files. Naïve Bayes, SMO and Random Tree classifiers are used for classification. The proposed work has achieved average accuracy of 78.8 %.


2020 ◽  
pp. 110-121
Author(s):  
Nataliia Verbych ◽  

Phonetics of Ukrainian dialects has long been and remains the object of analysis. Separate articles and thorough monographic studies confirm this. Researchers consider to describing the system-structural originality of speeches, establishing the composition of segmental units in different dialects, their significance in the phonological system of the dialect, systematizing knowledge about the manifestations and relationships of phonemes. Super segmental differences of Ukrainian dialects are insufficiently studied. Intonation is a collection of sound linguistic means that are used to express semantic, emotional, expressive and modal character of the phrase, communicative meaning and situational conditionality, stylistic color of the text and the individuality of dialect speakers’ expressive techniques. The value of intonation in organization of oral speech is determined by its function – segmentation, structuring and selection. The paper studies the intonation parameters that ensure the integrity of the text and perform the function of connecting its individual elements. The author described the super segmental organization of a dialect text, identified and explored prosodic means not only of a single word, phrase or phrase, described the relationship of these units within the text, taking into account its content and structure. The article focuses on the features of segmentation of the dialectal speech. This study shows the difference between real sound file and his fixing during an auditory analysis. Much attention is given to the difference between a syntax and real articulation of broadcasting. The study demonstrates that the intonation structure of the dialect text as a kind of spontaneous speech is peculiar. In dialect narratives, the relationship between syntax and pauses (as the most important markers of segmentation) is much more complex than in a read or pre-prepared text. In spontaneous dialect narratives, the correlation between content and form shifts due to the simultaneity and synchronicity of the processes of thinking, planning, and producing thought. In some parts of the text there is a violation of formal and grammatical connections, the boundaries of phrases / syntagmas are blurred, their prosodic design has no clear delimitative features, as in codified speech, which leads to variance in the division of the text into separate segments. Keywords: dialect narratives, intonation, contour, pitch, pause.


Bioacoustics ◽  
2017 ◽  
Vol 28 (1) ◽  
pp. 57-73 ◽  
Author(s):  
Marcelo Araya-Salas ◽  
Grace Smith-Vidaurre ◽  
Michael Webster

2013 ◽  
Vol 18 (1) ◽  
pp. 60-70 ◽  
Author(s):  
Elizabeth Hoffman

Adorno's theory of musical reproduction is unfinished, inconsistent and attuned only to score-based acoustic music – but it has relevance for electroacoustic performance as well. His theory prompts contemplation about what ‘good’ interpretation, and interpretation itself, means for fixed electroacoustic music. A digital sound file is frequently, if not typically, viewed as more rigid and precise than a score. This article uses Adorno's theory to compare ontologies of score and digital file realizations respectively, thus questioning the above assumption. Do electroacoustic works truly exist apart from their performed features, or is a given work only its performances? Different answers imply different work concepts and interpretive strategies. Toward the essay's goals, we examine three features often viewed as nonontological to an electroacoustic work, namely performed spatialisation, equalisation, and amplitude balance. We consider the impacts of these features when they are manipulated in real time, or performance to performance. As Adorno asks how choices of timing or dynamics dictate a notated work's aesthetic ‘clarity’, this paper asks how performed choices contribute to an electroacoustic work's clarity, and to the unique interpretive potential of electroacoustic music. Tape music and acousmatic music, with its diffusion tradition, are central to this paper's thesis; but multi-channel works are circumscribed by it as well.


Author(s):  
V. J Manzo

In this chapter, we will look at some of the ways that you can play back and record sound files. As you know, Max lets you design the way you control the variables in your patch. We will apply these design concepts to the ways we control the playback of recorded sound. We will also look at some ways to track the pitch of analog audio and convert it into MIDI numbers. By the end of this chapter, you will have written a program that allows you to play back sound files using a computer keyboard as a control interface as well as a program that tracks the pitch you’re singing from a microphone and automatically harmonizes in real time. We will create a simple patch that plays back some prerecorded files I have prepared. Please locate the8 “.aif ” audio files located in the Chapter 13 Examples folder. 1. Copy these 8 audio files to a new folder somewhere on your computer 2. In Max, create a new patch 3. Click File>Save As and save the patch as playing_sounds.maxpat in the same folder where you put these 8 audio files. There should be 9 files total in the folder (8 audio and 1 Max patch) 4. Close the patch playing_sounds.maxpat 5. Re-open playing_sounds.maxpat (the audio files will now be in the search path of the Max patch) We can play back the prerecorded audio files we just copied using an object called sfplay~. The sfplay~ object takes an argument to specify how many channels of audio you would like the object to handle. For instance, if you are loading a stereo (two channel) file, you can specify the argument 2. Loading a sound file is easy: simply send the sfplay~ object the message open. Playing back the sound is just as easy: send it a 1 or a 0 from a toggle. Let’s build a patch that plays back these files.


Author(s):  
V. J Manzo

In this chapter, we will examine some ways to interact with audio processing objects in formal compositions. Examples of traditional instrumentalists interacting with Max patches in concert performances are common. In the interest of copyright availability, we will examine a composition of mine for E♭clarinet and computer (a Max patch). The remaining example patches in this chapter will deal with audio processing as it relates to hearing and some aspects of perception. In this composition, discourse, the clarinetist plays from a score while the Max patch “listens” to the performer (using a microphone) and processes the clarinet sound in predetermined ways. The Max patch follows a time-based “score” of its own for performing the effects on the clarinet sound and, thus, processes the audio signal the same way each time the piece is performed. Our purpose in exploring this patch has less to do with the effects that are used or any aesthetic you get from the piece than with the implementation of a usable timeline that both the clarinetist and the computer can perform to. 1. Open the file discourse.maxpat from within the folder discourse located in the Chapter 20 Examples folder When the space bar is pressed, a clocker within the patch will begin triggering the events in the Max patch; this is like the score for the computer. These events assume that since the user has pressed the space bar, the patch can expect to hear the notes of the score played back at tempo to coincide with the different audio processing taking place within the patch. Unless you happen to have your E♭clarinet handy (a PDF of the score is also available in the discourse folder), we will use a demo sound file of a synthesized clarinet playing this piece in lieu of actually performing it. This will give us a sense of what the piece would sound like if we were to perform the clarinet part live.


Sign in / Sign up

Export Citation Format

Share Document