首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Investigation into the evolution of human language has involved evidence of many different kinds and approaches from many different disciplines. For full modern language, humans must have evolved a range of physical abilities for the production of our complex speech sounds, as well as sophisticated cognitive abilities. Human speech involves free‐flowing, intricately varied, rapid sound sequences suitable for the fast transfer of complex, highly flexible communication. Some aspects of human speech, such as our ability to manipulate the vocal tract to produce a wide range of different types of sounds that form vowels and consonants, have attracted considerable attention from those interested in the evolution of language. 1 , 2 However, one very important contributory skill, the human ability to attain very fine control of breathing during speech, has been neglected. Here we present evidence of the importance of breathing control to human speech, as well as evidence that our capabilities greatly exceed those of nonhuman primates. Human speech breathing demands fine neurological control of the respiratory muscles, integrated with cognitive processes and other factors. Evidence from comparison of the vertebral canals of fossil hominids and those of extant primates suggests that a major increase in thoracic innervation evolved in later hominid evolution, providing enhanced breathing control. If that is so, then earlier hominids would have had quite restricted speech patterns, whereas more recent hominids, with human‐like breath control abilities, would have been capable of faster, more varied speech sequences.  相似文献   

2.
Inferences on the evolution of human speech based on anatomical data must take into account its physiology, acoustics and perception. Human speech is generated by the supralaryngeal vocal tract (SVT) acting as an acoustic filter on noise sources generated by turbulent airflow and quasi-periodic phonation generated by the activity of the larynx. The formant frequencies, which are major determinants of phonetic quality, are the frequencies at which relative energy maxima will pass through the SVT filter. Neither the articulatory gestures of the tongue nor their acoustic consequences can be fractionated into oral and pharyngeal cavity components. Moreover, the acoustic cues that specify individual consonants and vowels are “encoded”, i.e., melded together. Formant frequency encoding makes human speech a vehicle for rapid vocal communication. Non-human primates lack the anatomy that enables modern humans to produce sounds that enhance this process, as well as the neural mechanisms necessary for the voluntary control of speech articulation. The specific claims of Duchin (1990) are discussed.  相似文献   

3.
4.
5.
Sound symbolism is a non-arbitrary relationship between speech sounds and meaning. We review evidence that, contrary to the traditional view in linguistics, sound symbolism is an important design feature of language, which affects online processing of language, and most importantly, language acquisition. We propose the sound symbolism bootstrapping hypothesis, claiming that (i) pre-verbal infants are sensitive to sound symbolism, due to a biologically endowed ability to map and integrate multi-modal input, (ii) sound symbolism helps infants gain referential insight for speech sounds, (iii) sound symbolism helps infants and toddlers associate speech sounds with their referents to establish a lexical representation and (iv) sound symbolism helps toddlers learn words by allowing them to focus on referents embedded in a complex scene, alleviating Quine''s problem. We further explore the possibility that sound symbolism is deeply related to language evolution, drawing the parallel between historical development of language across generations and ontogenetic development within individuals. Finally, we suggest that sound symbolism bootstrapping is a part of a more general phenomenon of bootstrapping by means of iconic representations, drawing on similarities and close behavioural links between sound symbolism and speech-accompanying iconic gesture.  相似文献   

6.
The transition from an amorphous subsong into mature song requires a series of vocal changes. By tracing song elements during development, we have shown that the imitation trajectory to the target could not be predicted based on monotonic progression of vocal changes, indicating an internal component that imposes constraints on song development. Here we further examine the nature of constraints on song imitation in the zebra finch. We first present techniques for identifying and tracing distinctive vocal changes, and then we examine how sequences of vocal change are expressed and coordinated. Examples suggest two types of constraints on song imitation, based on the nature of the temporal context. Developmentally diachronic constraints are imposed by sequential dependencies between vocal changes as a function of developmental time, whereas developmentally synchronic constraints are given by the acoustic context of notes within the song. Finally, we show that the tendency of birds to copy certain sounds in the song model before others might be related to such constraints. We suggest that documenting the full range of distinctive vocal changes and the coordination of their expression would be useful for testing mechanisms of vocal imitation.  相似文献   

7.
The sounds of human speech make human language a rapid medium of communication through a process of speech "encoding." The presence of sounds like the vowels [a], [i], and [u] makes this process possible. The supralaryngeal vocal tracts of newborn Homo sapiens and chimpanzee are similar and resemble the reconstructed vocal tract of the fossil La Chapelle-aux-Saints Neanderthal man. Vocal tract area functions that were directed toward making best possible approximations to the human vowels [a], [i], and [u], as well as certain consonantal configurations, were modeled by means of a computer program. The lack of these vowels in the phonetic repertories of these creatures, who lack a supralaryngeal pharyngeal region like that of adult Homo sapiens, may be concomitant with the absence of speech encoding and a consequently linguistic ability inferior to modern man.  相似文献   

8.
For many years the evolution of language has been seen as a disreputable topic, mired in fanciful “just so stories” about language origins. However, in the last decade a new synthesis of modern linguistics, cognitive neuroscience and neo-Darwinian evolutionary theory has begun to make important contributions to our understanding of the biology and evolution of language. I review some of this recent progress, focusing on the value of the comparative method, which uses data from animal species to draw inferences about language evolution. Discussing speech first, I show how data concerning a wide variety of species, from monkeys to birds, can increase our understanding of the anatomical and neural mechanisms underlying human spoken language, and how bird and whale song provide insights into the ultimate evolutionary function of language. I discuss the “descended larynx” of humans, a peculiar adaptation for speech that has received much attention in the past, which despite earlier claims is not uniquely human. Then I will turn to the neural mechanisms underlying spoken language, pointing out the difficulties animals apparently experience in perceiving hierarchical structure in sounds, and stressing the importance of vocal imitation in the evolution of a spoken language. Turning to ultimate function, I suggest that communication among kin (especially between parents and offspring) played a crucial but neglected role in driving language evolution. Finally, I briefly discuss phylogeny, discussing hypotheses that offer plausible routes to human language from a non-linguistic chimp-like ancestor. I conclude that comparative data from living animals will be key to developing a richer, more interdisciplinary understanding of our most distinctively human trait: language.  相似文献   

9.
The study aimed at investigation into the first step of language acquisition by Russian children and the imitation process as a possible mechanism of mastering the language through the sounds specific for their native language. On the basis of spectrographic, phonetic, and listener's analysis of the sounds infants, the beginning of shaping of the phonologic system in Russian language, was shown. It was shown that, during imitation of mother's voice, children acquire articulation skills and sound utterances approximating a phoneme category of the native language. The data obtained indicate that interaction between mother and child fulfills the function of learning transmitting a language data on the tongue trame.  相似文献   

10.
Mimicry is a classical example of adaptive signal design. Here, we review the current state of research into vocal mimicry in birds. Avian vocal mimicry is a conspicuous and often spectacular form of animal communication, occurring in many distantly related species. However, the proximate and ultimate causes of vocal mimicry are poorly understood. In the first part of this review, we argue that progress has been impeded by conceptual confusion over what constitutes vocal mimicry. We propose a modified version of Vane‐Wright's (1980) widely used definition of mimicry. According to our definition, a vocalisation is mimetic if the behaviour of the receiver changes after perceiving the acoustic resemblance between the mimic and the model, and the behavioural change confers a selective advantage on the mimic. Mimicry is therefore specifically a functional concept where the resemblance between heterospecific sounds is a target of selection. It is distinct from other forms of vocal resemblance including those that are the result of chance or common ancestry, and those that have emerged as a by‐product of other processes such as ecological convergence and selection for large song‐type repertoires. Thus, our definition provides a general and functionally coherent framework for determining what constitutes vocal mimicry, and takes account of the diversity of vocalisations that incorporate heterospecific sounds. In the second part we assess and revise hypotheses for the evolution of avian vocal mimicry in the light of our new definition. Most of the current evidence is anecdotal, but the diverse contexts and acoustic structures of putative vocal mimicry suggest that mimicry has multiple functions across and within species. There is strong experimental evidence that vocal mimicry can be deceptive, and can facilitate parasitic interactions. There is also increasing support for the use of vocal mimicry in predator defence, although the mechanisms are unclear. Less progress has been made in explaining why many birds incorporate heterospecific sounds into their sexual displays, and in determining whether these vocalisations are functionally mimetic or by‐products of sexual selection for other traits such as repertoire size. Overall, this discussion reveals a more central role for vocal mimicry in the behavioural ecology of birds than has previously been appreciated. The final part of this review identifies important areas for future research. Detailed empirical data are needed on individual species, including on the structure of mimetic signals, the contexts in which mimicry is produced, how mimicry is acquired, and the ecological relationships between mimic, model and receiver. At present, there is little information and no consensus about the various costs of vocal mimicry for the protagonists in the mimicry complex. The diversity and complexity of vocal mimicry in birds raises important questions for the study of animal communication and challenges our view of the nature of mimicry itself. Therefore, a better understanding of avian vocal mimicry is essential if we are to account fully for the diversity of animal signals.  相似文献   

11.

Background

Humans can easily restore a speech signal that is temporally masked by an interfering sound (e.g., a cough masking parts of a word in a conversation), and listeners have the illusion that the speech continues through the interfering sound. This perceptual restoration for human speech is affected by prior experience. Here we provide evidence for perceptual restoration in complex vocalizations of a songbird that are acquired by vocal learning in a similar way as humans learn their language.

Methodology/Principal Findings

European starlings were trained in a same/different paradigm to report salient differences between successive sounds. The birds'' response latency for discriminating between a stimulus pair is an indicator for the salience of the difference, and these latencies can be used to evaluate perceptual distances using multi-dimensional scaling. For familiar motifs the birds showed a large perceptual distance if discriminating between song motifs that were muted for brief time periods and complete motifs. If the muted periods were filled with noise, the perceptual distance was reduced. For unfamiliar motifs no such difference was observed.

Conclusions/Significance

The results suggest that starlings are able to perceptually restore partly masked sounds and, similarly to humans, rely on prior experience. They may be a suitable model to study the mechanism underlying experience-dependent perceptual restoration.  相似文献   

12.
Vocal mimicry by the Black-browed Reed Warbler Acrocephalus bistrigiceps was investigated. To identify mimicry objectively, we measured similarities between the sounds of models and those of Warblers by means of Principal Component Analysis (PCA) using a set of acoustic parameters. Of the sounds suspected of being mimicry according to visual inspection of sonagrams, only 57% were identifiable as mimicry according to PCA. Previous studies have not included quantitative criteria for assessing vocal mimicry, and our results suggest that judgements might not be reliable in the absence of objective criteria. Male Warblers incorporated the mimetic sounds into their songs, and each male mimicked 2–5 species. We found no evidence that females preferred males with large mimetic repertoires. This suggests that vocal mimicry has not evolved in response to selection by females in this species, although our analysis did not reveal entire mimetic repertoires in the Warbler songs.  相似文献   

13.
Research into speech perception by nonhuman animals can be crucially informative in assessing whether specific perceptual phenomena in humans have evolved to decode speech, or reflect more general traits. Birds share with humans not only the capacity to use complex vocalizations for communication but also many characteristics of its underlying developmental and mechanistic processes; thus, birds are a particularly interesting group for comparative study. This review first discusses commonalities between birds and humans in perception of speech sounds. Several psychoacoustic studies have shown striking parallels in seemingly speech-specific perceptual phenomena, such as categorical perception of voice-onset-time variation, categorization of consonants that lack phonetic invariance, and compensation for coarticulation. Such findings are often regarded as evidence for the idea that the objects of human speech perception are auditory or acoustic events rather than articulations. Next, I highlight recent research on the production side of avian communication that has revealed the existence of vocal tract filtering and articulation in bird species-specific vocalization, which has traditionally been considered a hallmark of human speech production. Together, findings in birds show that many of characteristics of human speech perception are not uniquely human but also that a comparative approach to the question of what are the objects of perception--articulatory or auditory events--requires careful consideration of species-specific vocal production mechanisms.  相似文献   

14.
The cortical regions involved in the different stages of speech production are relatively well-established, but their spatio-temporal dynamics remain poorly understood. In particular, the available studies have characterized neural events with respect to the onset of the stimulus triggering a verbal response. The core aspect of language production, however, is not perception but action. In this context, the most relevant question may not be how long after a stimulus brain events happen, but rather how long before the production act do they occur. We investigated speech production-related brain activity time-locked to vocal onset, in addition to the common stimulus-locked approach. We report the detailed temporal interplay between medial and left frontal activities occurring shortly before vocal onset. We interpret those as reflections of, respectively, word selection and word production processes. This medial-lateral organization is in line with that described in non-linguistic action control, suggesting that similar processes are at play in word production and non-linguistic action production. This novel view of the brain dynamics underlying word production provides a useful background for future investigations of the spatio-temporal brain dynamics that lead to the production of verbal responses.  相似文献   

15.
Although up to 25% of children with autism are non-verbal, there are very few interventions that can reliably produce significant improvements in speech output. Recently, a novel intervention called Auditory-Motor Mapping Training (AMMT) has been developed, which aims to promote speech production directly by training the association between sounds and articulatory actions using intonation and bimanual motor activities. AMMT capitalizes on the inherent musical strengths of children with autism, and offers activities that they intrinsically enjoy. It also engages and potentially stimulates a network of brain regions that may be dysfunctional in autism. Here, we report an initial efficacy study to provide 'proof of concept' for AMMT. Six non-verbal children with autism participated. Prior to treatment, the children had no intelligible words. They each received 40 individual sessions of AMMT 5 times per week, over an 8-week period. Probe assessments were conducted periodically during baseline, therapy, and follow-up sessions. After therapy, all children showed significant improvements in their ability to articulate words and phrases, with generalization to items that were not practiced during therapy sessions. Because these children had no or minimal vocal output prior to treatment, the acquisition of speech sounds and word approximations through AMMT represents a critical step in expressive language development in children with autism.  相似文献   

16.
A central challenge for articulatory speech synthesis is the simulation of realistic articulatory movements, which is critical for the generation of highly natural and intelligible speech. This includes modeling coarticulation, i.e., the context-dependent variation of the articulatory and acoustic realization of phonemes, especially of consonants. Here we propose a method to simulate the context-sensitive articulation of consonants in consonant-vowel syllables. To achieve this, the vocal tract target shape of a consonant in the context of a given vowel is derived as the weighted average of three measured and acoustically-optimized reference vocal tract shapes for that consonant in the context of the corner vowels /a/, /i/, and /u/. The weights are determined by mapping the target shape of the given context vowel into the vowel subspace spanned by the corner vowels. The model was applied for the synthesis of consonant-vowel syllables with the consonants /b/, /d/, /g/, /l/, /r/, /m/, /n/ in all combinations with the eight long German vowels. In a perception test, the mean recognition rate for the consonants in the isolated syllables was 82.4%. This demonstrates the potential of the approach for highly intelligible articulatory speech synthesis.  相似文献   

17.
To study the continuity in speech development, an investigation of vowel-like sounds recorded in the course of longitudinal research of speech formation was carried out in 12 infants beginning with the first month up to 12 months of their life. It was revealed that features ensuring separation of vowel-like sounds are specific by their amplitude relation and frequency location of the most expressed spectral maxima, including the maxima corresponding to the fundamental frequency. As fundamental frequency increased, the vowel-like sounds [a], [u], [i] began to exhibit specific changes of the amplitude ratios of the spectral maxima. It was established that peculiarities of frequency positions of spectral maxima and relations of their amplitudes may be the very features on the bases of which children compare their own vowel-like sounds to the vowels of the adults in sound imitation. These findings and literature data corroborated the opinion of continuity in speech development, beginning with the early preverbal vocalizations.  相似文献   

18.
We present evidence that the geographic context in which a language is spoken may directly impact its phonological form. We examined the geographic coordinates and elevations of 567 language locations represented in a worldwide phonetic database. Languages with phonemic ejective consonants were found to occur closer to inhabitable regions of high elevation, when contrasted to languages without this class of sounds. In addition, the mean and median elevations of the locations of languages with ejectives were found to be comparatively high. The patterns uncovered surface on all major world landmasses, and are not the result of the influence of particular language families. They reflect a significant and positive worldwide correlation between elevation and the likelihood that a language employs ejective phonemes. In addition to documenting this correlation in detail, we offer two plausible motivations for its existence. We suggest that ejective sounds might be facilitated at higher elevations due to the associated decrease in ambient air pressure, which reduces the physiological effort required for the compression of air in the pharyngeal cavity–a unique articulatory component of ejective sounds. In addition, we hypothesize that ejective sounds may help to mitigate rates of water vapor loss through exhaled air. These explications demonstrate how a reduction of ambient air density could promote the usage of ejective phonemes in a given language. Our results reveal the direct influence of a geographic factor on the basic sound inventories of human languages.  相似文献   

19.
Yamamoto K  Kawabata H 《PloS one》2011,6(12):e29414

Background

We ordinarily perceive our voice sound as occurring simultaneously with vocal production, but the sense of simultaneity in vocalization can be easily interrupted by delayed auditory feedback (DAF). DAF causes normal people to have difficulty speaking fluently but helps people with stuttering to improve speech fluency. However, the underlying temporal mechanism for integrating the motor production of voice and the auditory perception of vocal sound remains unclear. In this study, we investigated the temporal tuning mechanism integrating vocal sensory and voice sounds under DAF with an adaptation technique.

Methods and Findings

Participants produced a single voice sound repeatedly with specific delay times of DAF (0, 66, 133 ms) during three minutes to induce ‘Lag Adaptation’. They then judged the simultaneity between motor sensation and vocal sound given feedback. We found that lag adaptation induced a shift in simultaneity responses toward the adapted auditory delays. This indicates that the temporal tuning mechanism in vocalization can be temporally recalibrated after prolonged exposure to delayed vocal sounds. Furthermore, we found that the temporal recalibration in vocalization can be affected by averaging delay times in the adaptation phase.

Conclusions

These findings suggest vocalization is finely tuned by the temporal recalibration mechanism, which acutely monitors the integration of temporal delays between motor sensation and vocal sound.  相似文献   

20.
Gao S  Hu J  Gong D  Chen S  Kendrick KM  Yao D 《PloS one》2012,7(5):e38289
Consonants, unlike vowels, are thought to be speech specific and therefore no interactions would be expected between consonants and pitch, a basic element for musical tones. The present study used an electrophysiological approach to investigate whether, contrary to this view, there is integrative processing of consonants and pitch by measuring additivity of changes in the mismatch negativity (MMN) of evoked potentials. The MMN is elicited by discriminable variations occurring in a sequence of repetitive, homogeneous sounds. In the experiment, event-related potentials (ERPs) were recorded while participants heard frequently sung consonant-vowel syllables and rare stimuli deviating in either consonant identity only, pitch only, or in both dimensions. Every type of deviation elicited a reliable MMN. As expected, the two single-deviant MMNs had similar amplitudes, but that of the double-deviant MMN was also not significantly different from them. This absence of additivity in the double-deviant MMN suggests that consonant and pitch variations are processed, at least at a pre-attentive level, in an integrated rather than independent way. Domain-specificity of consonants may depend on higher-level processes in the hierarchy of speech perception.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号