Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
1,401
result(s) for
"Sound Spectrography"
Sort by:
USVSEG: A robust method for segmentation of ultrasonic vocalizations in rodents
2020
Rodents' ultrasonic vocalizations (USVs) provide useful information for assessing their social behaviors. Despite previous efforts in classifying subcategories of time-frequency patterns of USV syllables to study their functional relevance, methods for detecting vocal elements from continuously recorded data have remained sub-optimal. Here, we propose a novel procedure for detecting USV segments in continuous sound data containing background noise recorded during the observation of social behavior. The proposed procedure utilizes a stable version of the sound spectrogram and additional signal processing for better separation of vocal signals by reducing the variation of the background noise. Our procedure also provides precise time tracking of spectral peaks within each syllable. We demonstrated that this procedure can be applied to a variety of USVs obtained from several rodent species. Performance tests showed this method had greater accuracy in detecting USV syllables than conventional detection methods.
Journal Article
Inharmonic speech reveals the role of harmonicity in the cocktail party problem
2018
The “cocktail party problem” requires us to discern individual sound sources from mixtures of sources. The brain must use knowledge of natural sound regularities for this purpose. One much-discussed regularity is the tendency for frequencies to be harmonically related (integer multiples of a fundamental frequency). To test the role of harmonicity in real-world sound segregation, we developed speech analysis/synthesis tools to perturb the carrier frequencies of speech, disrupting harmonic frequency relations while maintaining the spectrotemporal envelope that determines phonemic content. We find that violations of harmonicity cause individual frequencies of speech to segregate from each other, impair the intelligibility of concurrent utterances despite leaving intelligibility of single utterances intact, and cause listeners to lose track of target talkers. However, additional segregation deficits result from replacing harmonic frequencies with noise (simulating whispering), suggesting additional grouping cues enabled by voiced speech excitation. Our results demonstrate acoustic grouping cues in real-world sound segregation.
Harmonicity is associated with a single sound source and may be a useful cue with which to segregate the speech of multiple talkers. Here the authors introduce a method for perturbing the constituent frequencies of speech and show that violating harmonicity degrades intelligibility of speech mixtures.
Journal Article
Revisiting the syntactic abilities of non-human animals: natural vocalizations and artificial grammar learning
by
ten Cate, Carel
,
Okanoya, Kazuo
in
Acoustic Stimulation - methods
,
Animal vocalization
,
Animals
2012
The domain of syntax is seen as the core of the language faculty and as the most critical difference between animal vocalizations and language. We review evidence from spontaneously produced vocalizations as well as from perceptual experiments using artificial grammars to analyse animal syntactic abilities, i.e. abilities to produce and perceive patterns following abstract rules. Animal vocalizations consist of vocal units (elements) that are combined in a species-specific way to create higher order strings that in turn can be produced in different patterns. While these patterns differ between species, they have in common that they are no more complex than a probabilistic finite-state grammar. Experiments on the perception of artificial grammars confirm that animals can generalize and categorize vocal strings based on phonetic features. They also demonstrate that animals can learn about the co-occurrence of elements or learn simple ‘rules’ like attending to reduplications of units. However, these experiments do not provide strong evidence for an ability to detect abstract rules or rules beyond finite-state grammars. Nevertheless, considering the rather limited number of experiments and the difficulty to design experiments that unequivocally demonstrate more complex rule learning, the question of what animals are able to do remains open.
Journal Article
Automatic Detection and Recognition of Pig Wasting Diseases Using Sound Data in Audio Surveillance Systems
by
Chung, Yongwha
,
Oh, Seunggeun
,
Lee, Jonguk
in
Accuracy
,
Animals
,
Auscultation - instrumentation
2013
Automatic detection of pig wasting diseases is an important issue in the management of group-housed pigs. Further, respiratory diseases are one of the main causes of mortality among pigs and loss of productivity in intensive pig farming. In this study, we propose an efficient data mining solution for the detection and recognition of pig wasting diseases using sound data in audio surveillance systems. In this method, we extract the Mel Frequency Cepstrum Coefficients (MFCC) from sound data with an automatic pig sound acquisition process, and use a hierarchical two-level structure: the Support Vector Data Description (SVDD) and the Sparse Representation Classifier (SRC) as an early anomaly detector and a respiratory disease classifier, respectively. Our experimental results show that this new method can be used to detect pig wasting diseases both economically (even a cheap microphone can be used) and accurately (94% detection and 91% classification accuracy), either as a standalone solution or to complement known methods to obtain a more accurate solution.
Journal Article
Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces
by
Savariaux, Christophe
,
Yvert, Blaise
,
Bocquelet, Florent
in
Acoustics
,
Aphasia
,
Biofeedback, Psychology - instrumentation
2016
Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer.
Journal Article
Directional Reflective Surface Formed via Gradient-Impeding Acoustic Meta-Surfaces
by
Song, Kyungjun
,
Lee, Seong-Hyun
,
Kim, Jedo
in
639/166/988
,
639/766/25/3927
,
Acoustic Stimulation
2016
Artificially designed acoustic meta-surfaces have the ability to manipulate sound energy to an extraordinary extent. Here, we report on a new type of directional reflective surface consisting of an array of sub-wavelength Helmholtz resonators with varying internal coiled path lengths, which induce a reflection phase gradient along a planar acoustic meta-surface. The acoustically reshaped reflective surface created by the gradient-impeding meta-surface yields a distinct focal line similar to a parabolic cylinder antenna, and is used for directive sound beamforming. Focused beam steering can be also obtained by repositioning the source (or receiver) off axis, i.e., displaced from the focal line. Besides flat reflective surfaces, complex surfaces such as convex or conformal shapes may be used for sound beamforming, thus facilitating easy application in sound reinforcement systems. Therefore, directional reflective surfaces have promising applications in fields such as acoustic imaging, sonic weaponry, and underwater communication.
Journal Article
Spatial and temporal patterns of sound production in East Greenland narwhals
by
Sinding, Mikkel H. S.
,
Blackwell, Susanna B.
,
Heide-Jørgensen, Mads Peter
in
Acoustic properties
,
Acoustics
,
Acoustics - instrumentation
2018
Changes in climate are rapidly modifying the Arctic environment. As a result, human activities-and the sounds they produce-are predicted to increase in remote areas of Greenland, such as those inhabited by the narwhals (Monodon monoceros) of East Greenland. Meanwhile, nothing is known about these whales' acoustic behavior or their reactions to anthropogenic sounds. This lack of knowledge was addressed by instrumenting six narwhals in Scoresby Sound (Aug 2013-2016) with Acousonde™ acoustic tags and satellite tags. Continuous recordings over up to seven days were used to describe the acoustic behavior of the whales, in particular their use of three types of sounds serving two different purposes: echolocation clicks and buzzes, which serve feeding, and calls, presumably used for social communication. Logistic regression models were used to assess the effects of location in time and space on buzzing and calling rates. Buzzes were mostly produced at depths of 350-650 m and buzzing rates were higher in one particular fjord, likely a preferred feeding area. Calls generally occurred at shallower depths (<100 m), with more than half of these calls occurring near the surface (<7 m), where the whales also spent more than half of their time. A period of silence following release, present in all subjects, was attributed to the capture and tagging operations, emphasizing the importance of longer (multi-day) records. This study provides basic life-history information on a poorly known species-and therefore control data in ongoing or future sound-effect studies.
Journal Article
Reconstruction of vocal interactions in a group of small songbirds
by
Hahnloser, Richard H R
,
Abramchuk, Andrei N
,
Vyssotski, Alexei L
in
631/158
,
631/1647/2198
,
631/378/2619
2014
Communications between animals such as zebra finches can be discriminated with back-attached acceleration recorders. In contrast to microphones, these devices record the carrier's signals only, allowing a more precise analysis of individual vocalizations during social interactions.
The main obstacle for investigating vocal interactions in vertebrates is the difficulty of discriminating individual vocalizations of rapidly moving, sometimes simultaneously vocalizing individuals. We developed a method of recording and analyzing individual vocalizations in free-ranging animals using ultraminiature back-attached sound and acceleration recorders. Our method allows the separation of zebra finch vocalizations irrespective of background noise and the number of vocalizing animals nearby.
Journal Article
Cough Sound Analysis Can Rapidly Diagnose Childhood Pneumonia
by
Triasih, Rina
,
Abeyratne, Udantha R.
,
Swarnkar, Vinayak
in
Adolescent
,
Algorithms
,
Asthma - diagnosis
2013
Pneumonia annually kills over 1,800,000 children throughout the world. The vast majority of these deaths occur in resource poor regions such as the sub-Saharan Africa and remote Asia. Prompt diagnosis and proper treatment are essential to prevent these unnecessary deaths. The reliable diagnosis of childhood pneumonia in remote regions is fraught with difficulties arising from the lack of field-deployable imaging and laboratory facilities as well as the scarcity of trained community healthcare workers. In this paper, we present a pioneering class of technology addressing both of these problems. Our approach is centred on the automated analysis of cough and respiratory sounds, collected
via
microphones that do not require physical contact with subjects. Cough is a cardinal symptom of pneumonia but the current clinical routines used in remote settings do not make use of coughs beyond
noting its existence
as a screening-in criterion. We hypothesized that cough carries vital information to diagnose pneumonia, and developed mathematical features and a pattern classifier system suited for the task. We collected cough sounds from 91 patients suspected of acute respiratory illness such as pneumonia, bronchiolitis and asthma. Non-contact microphones kept by the patient’s bedside were used for data acquisition. We extracted features such as non-Gaussianity and Mel Cepstra from cough sounds and used them to train a Logistic Regression classifier. We used the clinical diagnosis provided by the paediatric respiratory clinician as the gold standard to train and validate our classifier. The methods proposed in this paper could separate pneumonia from other diseases at a sensitivity and specificity of 94 and 75% respectively, based on parameters extracted from cough sounds alone. The inclusion of other simple measurements such as the presence of fever further increased the performance. These results show that cough sounds indeed carry critical information on the lower respiratory tract, and can be used to diagnose pneumonia. The performance of our method is far superior to those of existing WHO clinical algorithms for resource-poor regions. To the best of our knowledge, this is the first attempt in the world to diagnose pneumonia in humans using cough sound analysis. Our method has the potential to revolutionize the management of childhood pneumonia in remote regions of the world.
Journal Article
Use of Spectral/Cepstral Analyses for Differentiating Normal From Hypofunctional Voices in Sustained Vowel and Continuous Speech Contexts
2011
Purpose: In this study, the authors evaluated the diagnostic value of spectral/cepstral measures to differentiate dysphonic from nondysphonic voices using sustained vowels and continuous speech samples. Methodology: Thirty-two age- and gender-matched individuals (16 participants with dysphonia and 16 controls) were recorded reading a standard passage (The Rainbow Passage; Fairbanks, 1960) and sustaining the vowel /[alpha]/. Recorded voices were analyzed with custom software that calculated 4 spectral/cepstral measures. Results: Measures of cepstral peak prominence (CPP) and low-high spectral ratio (L/H ratio) were significantly different between groups in both speaking conditions; the standard deviation of the CPP was significantly different between groups in continuous speech only. In differentiating dysphonic individuals with a hypofunctional etiology from nondysphonic individuals, receiver operating characteristic (ROC) analyses demonstrated (a) high sensitivity and high specificity for the CPP in the sustained vowel condition and (b) high sensitivity and moderate specificity for the CPP in the speech condition. Conclusions: In a sample of dysphonic speakers (hypofunctional etiologies) versus typical speakers, spectral/cepstral measures of CPP and L/H ratio were able to differentiate these groups from one another in both vowel prolongation and continuous speech contexts with high sensitivity and specificity. The results of this study support the growing body of literature documenting the significant value of cepstral and other spectral-based acoustic measures to the clinical evaluation and management processes.
Journal Article