Ideophones, also termed mimetics or expressives, are marked words that depict sensory imagery. They are found in many of the world’s languages, and sizable lexical classes of ideophones are particularly well-documented in the languages of Asia, Africa, and the Americas. Ideophones are not limited to onomatopoeia like meow and smack but cover a wide range of sensory domains, such as manner of motion (e.g., plisti plasta ‘splish-splash’ in Basque), texture (e.g., tsaklii ‘rough’ in Ewe), and psychological states (e.g., wakuwaku ‘excited’ in Japanese). Across languages, ideophones stand out as marked words due to special phonotactics, expressive morphology including certain types of reduplication, and relative syntactic independence, in addition to production features like prosodic foregrounding and common co-occurrence with iconic gestures.
Three intertwined issues have been repeatedly debated in the century-long literature on ideophones. (a) Definition: Isolated descriptive traditions and cross-linguistic variation have sometimes obscured a typologically unified view of ideophones, but recent advances show the promise of a prototype definition of ideophones as conventionalized depictions in speech, with room for language-specific nuances. (b) Integration: The variable integration of ideophones across linguistic levels reveals an interaction between expressiveness and grammatical integration, and has important implications for how to conceive of dependencies between linguistic systems. (c) Iconicity: Ideophones form a natural laboratory for the study of iconic form-meaning associations in natural languages, and converging evidence from corpus and experimental studies suggests important developmental, evolutionary, and communicative advantages of ideophones.
Article
Ideophones (Mimetics, Expressives)
Kimi Akita and Mark Dingemanse
Article
Innateness of Language
Yarden Kedar
A fundamental question in epistemological philosophy is whether reason may be based on a priori knowledge—that is, knowledge that precedes and which is independent of experience. In modern science, the concept of innateness has been associated with particular behaviors and types of knowledge, which supposedly have been present in the organism since birth (in fact, since fertilization)—prior to any sensory experience with the environment.
This line of investigation has been traditionally linked to two general types of qualities: the first consists of instinctive and inflexible reflexes, traits, and behaviors, which are apparent in survival, mating, and rearing activities. The other relates to language and cognition, with certain concepts, ideas, propositions, and particular ways of mental computation suggested to be part of one’s biological make-up. While both these types of innatism have a long history (e.g., debate by Plato and Descartes), some bias appears to exist in favor of claims for inherent behavioral traits, which are typically accepted when satisfactory empirical evidence is provided. One famous example is Lorenz’s demonstration of imprinting, a natural phenomenon that obeys a predetermined mechanism and schedule (incubator-hatched goslings imprinted on Lorenz’s boots, the first moving object they encountered). Likewise, there seems to be little controversy in regard to predetermined ways of organizing sensory information, as is the case with the detection and classification of shapes and colors by the mind.
In contrast, the idea that certain types of abstract knowledge may be part of an organism’s biological endowment (i.e., not learned) is typically met with a greater sense of skepticism. The most influential and controversial claim for such innate knowledge in modern science is Chomsky’s nativist theory of Universal Grammar in language, which aims to define the extent to which human languages can vary; and the famous Argument from the Poverty of the Stimulus. The main Chomskyan hypothesis is that all human beings share a preprogrammed linguistic infrastructure consisting of a finite set of general principles, which can generate (through combination or transformation) an infinite number of (only) grammatical sentences. Thus, the innate grammatical system constrains and structures the acquisition and use of all natural languages.
Article
Issues in Neurolinguistic Studies of Morphology
Robert Fiorentino
Research in neurolinguistics examines how language is organized and processed in the human brain. The findings from neurolinguistic studies on language can inform our understanding of the basic ingredients of language and the operations they undergo. In the domain of the lexicon, a major debate concerns whether and to what extent the morpheme serves as a basic unit of linguistic representation, and in turn whether and under what circumstances the processing of morphologically complex words involves operations that identify, activate, and combine morpheme-level representations during lexical processing. Alternative models positing some role for morphemes argue that complex words are processed via morphological decomposition and composition in the general case (full-decomposition models), or only under certain circumstances (dual-route models), while other models do not posit a role for morphemes (non-morphological models), instead arguing that complex words are related to their constituents not via morphological identity, but either via associations among whole-word representations or via similarity in formal and/or semantic features. Two main approaches to investigating the role of morphemes from a neurolinguistic perspective are neuropsychology, in which complex word processing is typically investigated in cases of brain insult or neurodegenerative disease, and brain imaging, which makes it possible to examine the temporal dynamics and neuroanatomy of complex word processing as it occurs in the brain. Neurolinguistic studies on morphology have examined whether the processing of complex words involves brain mechanisms that rapidly segment the input into potential morpheme constituents, how and under what circumstances morpheme representations are accessed from the lexicon, and how morphemes are combined to form complex morphosyntactic and morpho-semantic representations. Findings from this literature broadly converge in suggesting a role for morphemes in complex word processing, although questions remain regarding the precise time course by which morphemes are activated, the extent to which morpheme access is constrained by semantic or form properties, as well as regarding the brain mechanisms by which morphemes are ultimately combined into complex representations.
Article
Iterative Meaning in Mandarin Chinese: Linguistic Factors and its Processing
Huei-ling Lai and Yao-Ying Lai
Sentential meaning that emerges compositionally is not always transparent as one-to-one mapping from syntactic structure to semantic representation; oftentimes, the meaning is underspecified (morphosyntactically unsupported), not explicitly conveyed via overt linguistic devices. Compositional meaning is obtained during comprehension. The associated issues are explored by examining linguistic factors that modulate the construal of underspecified iterative meaning in Mandarin Chinese (MC). In this case, the factors include lexical aspect of verbs, the interval-lengths denoted by post-verbal durative adverbials, and boundary specificity denoted by preverbal versus post-verbal temporal adverbials. The composition of a punctual verb (e.g., jump, poke) with a durative temporal adverbial like Zhangsan tiao-le shi fenzhong. Zhangsan jump-LE ten minute ‘Zhangsan jumped for ten minutes’ engenders an iterative meaning, which is morphosyntactically absent yet fully understood by comprehenders. Contrastively, the counterpart involving a durative verb (e.g., run, swim) like Zhangsan pao-le shi fenzhong Zhangsan run-LE ten minute ‘Zhangsan ran for ten minutes’ engenders a continuous reading with identical syntactic structure. Psycholinguistically, processing such underspecified meaning in real time has been shown to require greater effort than the transparent counterpart. This phenomenon has been attested cross-linguistically; yet how it is manifested in MC, a tenseless language, remains understudied. In addition, durative temporal adverbials like yizhi/buduandi ‘continuously,’ which appear preverbally in MC, also engender an iterative meaning when composed with a punctual verb like Zhangsan yizhi/buduandi tiao. Zhangsan continuously jump ‘Zhangsan jumped continuously.’ Crucially, unlike the post-verbal adverbials that encode specific boundaries for the denoted intervals, these preverbal adverbials refer to continuous time spans without specific endpoints. The difference in boundary specificity between the two adverbial types, while both being durative, is hypothesized to modulate the processing profiles of aspectual comprehension.
Results of the online (timed) questionnaire showed (a) an effect of boundary specificity: sentences with post-verbal adverbials that encode [+specific boundary] were rated lower in the naturalness-rating task and induced longer response time (RT) in iterativity judgements, as compared to preverbal adverbials that encode [−specific boundary]; (b) in composition with post-verbal adverbials that are [+specific boundary], sentences involving durative verbs elicited lower rating scores and longer RT of iterativity judgements than the counterpart involving punctual verbs. These suggest that the comprehension of underspecified iterative meaning is modulated by both cross-linguistically similar parameters and language-specific systems of temporal reference, by which MC exhibits a typological difference in processing profiles. Overall, the patterns are consistent with the Context-Dependence approach to semantic underspecification: comprehenders compute the ultimate reading (iterative versus continuous) by taking both the sentential and extra-sentential information into consideration in a given context.
Article
Japanese Psycholinguistics
Mineharu Nakayama
The Japanese psycholinguistics research field is moving rapidly in many different directions as it includes various sub-linguistics fields (e.g., phonetics/phonology, syntax, semantics, pragmatics, discourse studies). Naturally, diverse studies have reported intriguing findings that shed light on our language mechanism. This article presents a brief overview of some of the notable early 21st century studies mainly from the language acquisition and processing perspectives. The topics are divided into various sections: the sound system, the script forms, reading and writing, morpho-syntactic studies, word and sentential meanings, and pragmatics and discourse studies sections. Studies on special populations are also mentioned.
Studies on the Japanese sound system have advanced our understanding of L1 and L2 (first and second language) acquisition and processing. For instance, more evidence is provided that infants form adult-like phonological grammar by 14 months in L1, and disassociation of prosody is reported from one’s comprehension in L2. Various cognitive factors as well as L1 influence the L2 acquisition process. As the Japanese language users employ three script forms (hiragana, katakana, and kanji) in a single sentence, orthographic processing research reveal multiple pathways to process information and the influence of memory. Adult script decoding and lexical processing has been well studied and research data from special populations further helps us to understand our vision-to-language mapping mechanism. Morpho-syntactic and semantic studies include a long debate on the nativist (generative) and statistical learning approaches in L1 acquisition. In particular, inflectional morphology and quantificational scope interaction in L1 acquisition bring pros and cons of both approaches as a single approach. Investigating processing mechanisms means studying cognitive/perceptual devices. Relative clause processing has been well-discussed in Japanese because Japanese has a different word order (SOV) from English (SVO), allows unpronounced pronouns and pre-verbal word permutations, and has no relative clause marking at the verbal ending (i.e., morphologically the same as the matrix ending). Behavioral and neurolinguistic data increasingly support incremental processing like SVO languages and an expectancy-driven processor in our L1 brain. L2 processing, however, requires more study to uncover its mechanism, as the literature is scarce in both L2 English by Japanese speakers and L2 Japanese by non-Japanese speakers. Pragmatic and discourse processing is also an area that needs to be explored further. Despite the typological difference between English and Japanese, the studies cited here indicate that our acquisition and processing devices seem to adjust locally while maintaining the universal mechanism.
Article
Korean Syntax
James Hye Suk Yoon
The syntax of Korean is characterized by several signature properties. One signature property is head-finality. Word order variations and restrictions obey head-finality. Korean also possesses wh in-situ as well as internally headed relative clauses, as is typical of a head-final language. Another major signature property is dependent-marking. Korean has systematic case-marking on nominal dependents and very little, if any, head-marking. Case-marking and related issues, such as multiple case constructions, case alternations, case stacking, case-marker ellipsis, and case-marking on adjuncts, are front and center properties of Korean syntax as viewed from the dependent-marking perspective. Research on these aspects of Korean has contributed to the theoretical understanding of case and grammatical relations in linguistic theory. Korean is also characterized by agglutinative morphosyntax. Many issues in Korean syntax straddle the morphology-syntax boundary. Korean morphosyntax constitutes a fertile testing ground for ongoing debates about the relationship between morphology and syntax in domains such as coordination, deverbal nominalizations (mixed category constructions), copula, and other denominal constructions. Head-finality and agglutinative morphosyntax intersect in domains such as complex/serial verb and auxiliary verb constructions. Negation, which is a type of auxiliary verb construction, and the related phenomena of negative polarity licensing, offer important evidence for crosslinguistic understanding of these phenomena. Finally, there is an aspect of Korean syntax that reflects areal contact. Lexical and grammatical borrowing, topic prominence, pervasive occurrence of null arguments and ellipsis, as well as a complex system of anaphoric expressions, resulted from sustained contact with neighboring Sino-Tibetan languages.
Article
Language Revitalization
Aidan Pine and Mark Turin
The world is home to an extraordinary level of linguistic diversity, with roughly 7,000 languages currently spoken and signed. Yet this diversity is highly unstable and is being rapidly eroded through a series of complex and interrelated processes that result in or lead to language loss. The combination of monolingualism and networks of global trade languages that are increasingly technologized have led to over half of the world’s population speaking one of only 13 languages. Such linguistic homogenization leaves in its wake a linguistic landscape that is increasingly endangered.
A wide range of factors contribute to language loss and attrition. While some—such as natural disasters—are unique to particular language communities and specific geographical regions, many have similar origins and are common across endangered language communities around the globe. The harmful legacy of colonization and the enduring impact of disenfranchising policies relating to Indigenous and minority languages are at the heart of language attrition from New Zealand to Hawai’i, and from Canada to Nepal.
Language loss does not occur in isolation, nor is it inevitable or in any way “natural.” The process also has wide-ranging social and economic repercussions for the language communities in question. Language is so heavily intertwined with cultural knowledge and political identity that speech forms often serve as meaningful indicators of a community’s vitality and social well-being. More than ever before, there are vigorous and collaborative efforts underway to reverse the trend of language loss and to reclaim and revitalize endangered languages. Such approaches vary significantly, from making use of digital technologies in order to engage individual and younger learners to community-oriented language nests and immersion programs. Drawing on diverse techniques and communities, the question of measuring the success of language revitalization programs has driven research forward in the areas of statistical assessments of linguistic diversity, endangerment, and vulnerability. Current efforts are re-evaluating the established triad of documentation-conservation-revitalization in favor of more unified, holistic, and community-led approaches.
Article
Languages of the World
William R. Leben
About 7,000 languages are spoken around the world today. The actual number depends on where the line is drawn between language and dialect—an arbitrary decision, because languages are always in flux. But specialists applying a reasonably uniform criterion across the globe count well over 2,000 languages in Asia and Africa, while Europe has just shy of 300. In between are the Pacific region, with over 1,300 languages, and the Americas, with just over 1,000. Languages spoken natively by over a million speakers number around 250, but the vast majority have very few speakers. Something like half are thought likely to disappear over the next few decades, as speakers of endangered languages turn to more widely spoken ones.
The languages of the world are grouped into perhaps 430 language families, based on their origin, as determined by comparing similarities among languages and deducing how they evolved from earlier ones. As with languages, there’s quite a lot of disagreement about the number of language families, reflecting our meager knowledge of many present-day languages and even sparser knowledge of their history. The figure 430 comes from Glottolog.org, which actually lists them all. While the world’s language families may well go back to a smaller number of original languages, even to a single mother tongue, scholars disagree on how far back current methods permit us to trace the history of languages.
While it is normal for languages to borrow from other languages, occasionally a totally new language is created by mixing elements of two distinct languages to such a degree that we would not want to identify one of the source languages as the mother tongue. This is what led to the development of Media Lengua, a language of Ecuador formed through contact among speakers of Spanish and speakers of Quechua. In this language, practically all the word stems are from Spanish, while all of the endings are from Quechua. Just a handful of languages have come into being in this way, but less extreme forms of language mixture have resulted in over a hundred pidgins and creoles currently spoken in many parts of the world. Most arose during Europe’s colonial era, when European colonists used their language to communicate with local inhabitants, who in turn blended vocabulary from the European language with grammar largely from their native language.
Also among the languages of the world are about 300 sign languages used mainly in communicating among and with the deaf. The structure of sign languages typically has little historical connection to the structure of nearby spoken languages.
Some languages have been constructed expressly, often by a single individual, to meet communication demands among speakers with no common language. Esperanto, designed to serve as a universal language and used as a second language by some two million, according to some estimates, is the prime example, but it is only one among several hundred would-be international auxiliary languages.
This essay surveys the languages of the world continent by continent, ending with descriptions of sign languages and of pidgins and creoles. A set of references grouped by section appears at the very end. The main source for data on language classification, numbers of languages, and speakers is the 19th edition of Ethnologue (see Resources), except where a different source is cited.
Article
Learning and Using Morphology and Morphosyntax in a Second Language
Laurie Beth Feldman and Judith F. Kroll
We summarize findings from across a range of methods, including behavioral measures of overall processing speed and accuracy, electrophysiological indices that tap into the early time course of language processing, and neural measures using structural and functional imaging. We argue that traditional claims about rigid constraints on the ability of late bilinguals to exploit the meaning and form of the morphology and morphosyntax in a second language should be revised so as to move away from all or none command of structures motivated from strict dichotomies among linguistic categories of morphology. We describe how the dynamics of morphological processing in neither monolingual or bilingual speakers is easily characterized in terms of the potential to decompose words into their constituent morphemes and that morphosyntactic processing is not easily characterized in terms of categories of structures that are learnable and those that are unlearnable by bilingual and nonnative speakers. Instead, we emphasize the high degree of variability across individuals and plasticity within individuals in their ability to successfully learn and use even subtle aspects of a second language. Further, both of the bilingual’s two languages become active when even one language is engaged, and parallel activation has consequences that shape both languages, thus their influence is not in the unidirectional manner that was traditionally assumed. We briefly discuss the nature of possible constraints and directions for future research.
Article
Lexical Acquisition and the Structure of the Mental Lexicon
Eve V. Clark
The words and word-parts children acquire at different stages offer insights into how the mental lexicon might be organized. Children first identify ‘words,’ recurring sequences of sounds, in the speech stream, attach some meaning to them, and, later, analyze such words further into parts, namely stems and affixes. These are the elements they store in memory in order to recognize them on subsequent occasions. They also serve as target models when children try to produce those words themselves. When they coin words, they make use of bare stems, combine certain stems with each other, and sometimes add affixes as well. The options they choose depend on how much they need to add to coin a new word, which familiar elements they can draw on, and how productive that option is in the language. Children’s uses of stems and affixes in coining new words also reveal that they must be relying on one representation in comprehension and a different representation in production. For comprehension, they need to store information about the acoustic properties of a word, taking into account different occasions, different speakers, and different dialects, not to mention second-language speakers. For production, they need to work out which articulatory plan to follow in order to reproduce the target word. And they take time to get their production of a word aligned with the representation they have stored for comprehension. In fact, there is a general asymmetry here, with comprehension being ahead of production for children, and also being far more extensive than production, for both children and adults. Finally, as children add more words to their repertoires, they organize and reorganize their vocabulary into semantic domains. In doing this, they make use of pragmatic directions from adults that help them link related words through a variety of semantic relations.
Article
Lexical Representations in Language Processing
Gary Libben
Words are the backbone of language activity. An average 20-year-old native speaker of English will have a vocabulary of about 42,000 words. These words are connected with one another within the larger network of lexical knowledge that is termed the mental lexicon. The metaphor of a mental lexicon has played a central role in the development of theories of language and mind and has provided an intellectual meeting ground for psychologists, neurolinguists, and psycholinguists. Research on the mental lexicon has shown that lexical knowledge is not static. New words are acquired throughout the life span, creating very large increases in the richness of connectivity within the lexical system and changing the system as a whole. Because most people in the world speak more than one language, the default mental lexicon may be a multilingual one. Such a mental lexicon differs substantially from a lexicon of an individual language and would lead to the creation of new integrated lexical systems due to the pressure on the system to organize and access lexical knowledge in a homogenous manner. The mental lexicon contains both word knowledge and morphological knowledge. There is also evidence that it contains multiword strings such as idioms and lexical bundles. This speaks in support of a nonrestrictive “big tent” view of units of representation within the mental lexicon. Changes in research on lexical representations in language processing have emphasized lexical action and the role of learning. Although the metaphor of words as distinct representations within a lexical store has served to advance knowledge, it is more likely that words are best seen as networks of activity that are formed and affected by experience and learning throughout the life span.
Article
Locality in Syntax
Adriana Belletti
Phenomena involving the displacement of syntactic units are widespread in human languages. The term displacement refers here to a dependency relation whereby a given syntactic constituent is interpreted simultaneously in two different positions. Only one position is pronounced, in general the hierarchically higher one in the syntactic structure. Consider a wh-question like (1) in English:
(1) Whom did you give the book to
The phrase containing the interrogative wh-word is located at the beginning of the clause, and this guarantees that the clause is interpreted as a question about this phrase; at the same time, whom is interpreted as part of the argument structure of the verb give (the copy, in <> brackets). In current terms, inspired by minimalist developments in generative syntax, the phrase whom is first merged as (one of) the complement(s) of give (External Merge) and then re-merged (Internal Merge, i.e., movement) in the appropriate position in the left periphery of the clause. This peripheral area of the clause hosts operator-type constituents, among which interrogative ones (yielding the relevant interpretation: for which x, you gave a book to x, for sentence 1). Scope-discourse phenomena—such as, e.g., the raising of a question as in (1), the focalization of one constituent as in TO JOHN I gave the book (not to Mary)—have the effect that an argument of the verb is fronted in the left periphery of the clause rather than filling its clause internal complement position, whence the term displacement. Displacement can be to a position relatively close to the one of first merge (the copy), or else it can be to a position farther away. In the latter case, the relevant dependency becomes more long-distance than in (1), as in (2)a and even more so (2)b:
(2)
a Whom did Mary expect [that you would give the book to]
b Whom do you think [that Mary expected [that you would give the book to ]]
50 years or so of investigation on locality in formal generative syntax have shown that, despite its potentially very distant realization, syntactic displacement is in fact a local process. The audible position in which a moved constituent is pronounced and the position of its copy inside the clause can be far from each other. However, the long-distance dependency is split into steps through iterated applications of short movements, so that any dependency holding between two occurrences of the same constituent is in fact very local. Furthermore, there are syntactic domains that resist movement out of them, traditionally referred to as islands. Locality is a core concept of syntactic computations. Syntactic locality requires that syntactic computations apply within small domains (cyclic
domains), possibly in the mentioned iterated way (successive cyclicity), currently rethought of in terms of Phase theory. Furthermore, in the Relativized Minimality tradition, syntactic locality requires that, given X . . . Z . . . Y, the dependency between the relevant constituent in its target position X and its first merge position Y should not be interrupted by any constituent Z which is similar to X in relevant formal features and thus intervenes, blocking the relation between X and Y. Intervention locality has also been shown to allow for an explicit characterization of aspects of children’s linguistic development in their capacity to compute complex object dependencies (also relevant in different impaired populations).
Article
Morphological Units: A Theoretical and Psycholinguistic Perspective
Dominiek Sandra
Speakers can transfer meanings to each other because they represent them in a perceptible form. Phonology and syntactic structure are two levels of linguistic form. Morphemes are situated in-between them. Like phonemes they have a phonological component, and like syntactic structures they carry relational information. A distinction can be made between inflectional and lexical morphology. Both are devices in the service of communicative efficiency, by highlighting grammatical and semantic relations, respectively. Morphological structure has also been studied in psycholinguistics, especially by researchers who are interested in the process of visual word recognition. They found that a word is recognized more easily when it belongs to a large morphological family, which suggests that the mental lexicon is structured along morphological lines. The semantic transparency of a word’s morphological structure plays an important role. Several findings also suggest that morphology plays an important role at a pre-lexical processing level as well. It seems that morphologically complex words are subjected to a process of blind morphological decomposition before lexical access is attempted.
Article
Neurolinguistic Approaches in Morphology
Niels O. Schiller
Neurolinguistic approaches to morphology include the main theories of morphological representation and processing in the human mind, such as full-listing, full-parsing, and hybrid dual-route models, and how the experimental evidence that has been acquired to support these theories uses different neurolinguistic paradigms (visual and auditory priming, violation, long-lag priming, picture-word interference, etc.) and methods (electroencephalography [EEG]/event-related brain potential [ERP], functional magnetic resonance imaging [fMRI], neuropsychology, and so forth).
Article
Neurolinguistic Research on the Romance Languages
Valentina Bambini and Paolo Canal
Neurolinguistics is devoted to the study of the language-brain relationship, using the methodologies of neuropsychology and cognitive neuroscience to investigate how linguistic categories are grounded in the brain. Although the brain infrastructure for language is invariable across cultures, neural networks might operate differently depending on language-specific features. In this respect, neurolinguistic research on the Romance languages, mostly French, Italian, and Spanish, proved key to progress the field, especially with specific reference to how the neural infrastructure for language works in the case of more richly inflected systems than English.
Among the most popular domains of investigation are agreement patterns, where studies on Spanish and Italian showed that agreement across features and domains (e.g., number or gender agreement) engages partially different neural substrates. Also, studies measuring the electrophysiological response suggested that agreement processing is a composite mechanism involving different temporal steps. Another domain is the noun-verb distinction, where studies on the Romance languages indicated that the brain is more sensitive to the greater morphosyntactic engagement of verbs compared with nouns rather than to the grammatical class distinction per se.
Concerning language disorders, the Romance languages shed new light on inflectional errors in aphasic speakers and contributed to revise the notion of agrammatism, which is not simply omission of morphemes but might involve incorrect substitution from the inflectional paradigm. Also, research in the Romance domain showed variation in degree and pattern of reading impairments due to language-specific segmental and suprasegmental features.
Despite these important contributions, the Romance family, with its multitude of languages and dialects and a richly documented diachronic evolution, is a still underutilized ‘treasure house’ for neurolinguistic research, with significant room for investigations exploring the brain signatures of language variation in time and space and refining the linking between linguistic categories and neurobiological primitives.
Article
New Computational Methods and the Study of the Romance Languages
Basilio Calderone and Vito Pirrelli
Nowadays, computer models of human language are instrumental to millions of people, who use them every day with little if any awareness of their existence and role. Their exponential development has had a huge impact on daily life through practical applications like machine translation or automated dialogue systems. It has also deeply affected the way we think about language as an object of scientific inquiry. Computer modeling of Romance languages has helped scholars develop new theoretical frameworks and new ways of looking at traditional approaches. In particular, computer modeling of lexical phenomena has had a profound influence on some fundamental issues in human language processing, such as the purported dichotomy between rules and exceptions, or grammar and lexicon, the inherently probabilistic nature of speakers’ perception of analogy and word internal structure, and their ability to generalize to novel items from attested evidence. Although it is probably premature to anticipate and assess the prospects of these models, their current impact on language research can hardly be overestimated. In a few years, data-driven assessment of theoretical models is expected to play an irreplaceable role in pacing progress in all branches of language sciences, from typological and pragmatic approaches to cognitive and formal ones.
Article
Number in Language
Paolo Acquaviva
Number is the category through which languages express information about the individuality, numerosity, and part structure of what we speak about. As a linguistic category it has a morphological, a morphosyntactic, and a semantic dimension, which are variously interrelated across language systems. Number marking can apply to a more or less restricted part of the lexicon of a language, being most likely on personal pronouns and human/animate nouns, and least on inanimate nouns. In the core contrast, number allows languages to refer to ‘many’ through the description of ‘one’; the sets referred to consist of tokens of the same type, but also of similar types, or of elements pragmatically associated with one named individual. In other cases, number opposes a reading of ‘one’ to a reading as ‘not one,’ which includes masses; when the ‘one’ reading is morphologically derived from the ‘not one,’ it is called a singulative. It is rare for a language to have no linguistic number at all, since a ‘one–many’ opposition is typically implied at least in pronouns, where the category of person discriminates the speaker as ‘one.’ Beyond pronouns, number is typically a property of nouns and/or determiners, although it can appear on other word classes by agreement. Verbs can also express part-structural properties of events, but this ‘verbal number’ is not isomorphic to nominal number marking. Many languages allow a variable proportion of their nominals to appear in a ‘general’ form, which expresses no number information. The main values of number-marked elements are singular and plural; dual and a much rarer trial also exist. Many languages also distinguish forms interpreted as paucals or as greater plurals, respectively, for small and usually cohesive groups and for generically large ones. A broad range of exponence patterns can express these contrasts, depending on the morphological profile of a language, from word inflections to freestanding or clitic forms; certain choices of classifiers also express readings that can be described as ‘plural,’ at least in certain interpretations. Classifiers can co-occur with other plurality markers, but not when these are obligatory as expressions of an inflectional paradigm, although this is debated, partly because the notion of classifier itself subsumes distinct phenomena. Many languages, especially those with classifiers, encode number not as an inflectional category, but through word-formation operations that express readings associated with plurality, including large size. Current research on number concerns all its morphological, morphosyntactic, and semantic dimensions, in particular the interrelations of them as part of the study of natural language typology and of the formal analysis of nominal phrases. The grammatical and semantic function of number and plurality are particularly prominent in formal semantics and in syntactic theory.
Article
Paradigms in Morphology
Petar Milin and James P. Blevins
Studies of the structure and function of paradigms are as old as the Western grammatical tradition. The central role accorded to paradigms in traditional approaches largely reflects the fact that paradigms exhibit systematic patterns of interdependence that facilitate processes of analogical generalization. The recent resurgence of interest in word-based models of morphological processing and morphological structure more generally has provoked a renewed interest in paradigmatic dimensions of linguistic structure. Current methods for operationalizing paradigmatic relations and determining the behavioral correlates of these relations extend paradigmatic models beyond their traditional boundaries. The integrated perspective that emerges from this work is one in which variation at the level of individual words is not meaningful in isolation, but rather guides the association of words to paradigmatic contexts that play a role in their interpretation.
Article
Penutian Languages
Anthony P. Grant
The Penutian language family, Penutian phylum, or better still, Penutian hypothesis is one of the largest genealogical linguistic groupings to have been proposed for western North America. It involves 16 families or isolates. Only a few of these families are demonstrably relatable to one another according to current knowledge and diachronic techniques. Sometimes Penutian is split by observers into groups of languages assumed to be interrelated, and this is done without assumptions that the groups themselves are interrelated.
This article focuses on the Canadian and US languages in “Sapir’s Penutian,” the most commonly accepted version; the most southerly family within Penutian is thus held as Yokutsan of California’s Sierra Nevada. It discusses the subclassification of the so-called Penutian languages into families and smaller units; aspects of their phonology, morphosyntax, and contact histories; and issues in their revitalization and the potential reconstruction of Proto-Penutian.
Article
The Phonetics of Prosody
Amalia Arvaniti
Prosody is an umbrella term used to cover a variety of interconnected and interacting phenomena, namely stress, rhythm, phrasing, and intonation. The phonetic expression of prosody relies on a number of parameters, including duration, amplitude, and fundamental frequency (F0). The same parameters are also used to encode lexical contrasts (such as tone), as well as paralinguistic phenomena (such as anger, boredom, and excitement). Further, the exact function and organization of the phonetic parameters used for prosody differ across languages. These considerations make it imperative to distinguish the linguistic phenomena that make up prosody from their phonetic exponents, and similarly to distinguish between the linguistic and paralinguistic uses of the latter. A comprehensive understanding of prosody relies on the idea that speech is prosodically organized into phrasal constituents, the edges of which are phonetically marked in a number of ways, for example, by articulatory strengthening in the beginning and lengthening at the end. Phrases are also internally organized either by stress, that is around syllables that are more salient relative to others (as in English and Spanish), or by the repetition of a relatively stable tonal pattern over short phrases (as in Korean, Japanese, and French). Both types of organization give rise to rhythm, the perception of speech as consisting of groups of a similar and repetitive pattern. Tonal specification over phrases is also used for intonation purposes, that is, to mark phrasal boundaries, and express information structure and pragmatic meaning. Taken together, the components of prosody help with the organization and planning of speech, while prosodic cues are used by listeners during both language acquisition and speech processing. Importantly, prosody does not operate independently of segments; rather, it profoundly affects segment realization, making the incorporation of an understanding of prosody into experimental design essential for most phonetic research.