Analogy is traditionally regarded as one of the three main factors responsible for language change, along with sound change and borrowing. Whereas sound change is understood to be phonetically motivated and blind to structural patterns and semantic and functional relationships, analogy is licensed precisely by those patterns and relationships. In the Neogrammarian tradition, analogical change is regarded, at least largely, as a by-product of the normal operation (acquisition, representation, and use) of the mental grammar. Historical linguists commonly use proportional equations of the form A : B = C : X to represent analogical innovations, where A, B, and C are (sets of) word forms known to the innovator, who solves for X by discerning a formal relationship between A and B and then deductively arriving at a form that is related to C in the same way that B is related to A.
Along with the core type of analogical change captured by proportional equations, most historical linguists include a number of other phenomena under the analogy umbrella. Some of these, such as paradigm leveling—the reduction or elimination of stem alternations in paradigms—are arguably largely proportional, but others such as contamination and folk etymology seem to have less to do with the normal operation of the mental grammar and instead involve some kind of interference among the mental representations of phonetically or semantically similar forms.
The Neogrammarian approach to analogical change has been criticized and challenged on a variety of grounds, and a number of important scholars use the term “analogy” in a rather different sense, to refer to the role that phonological and/or semantic similarity play in the influence that forms exert on each other.
21-40 of 475 Results
Article
K. A. Jayaseelan
The Dravidian languages have a long-distance reflexive anaphor taan
. (It is taan
in Tamil and Malayalam, taanu in Kannada and tanu in Telugu.) As is the case with other long-distance anaphors, it is subject-oriented; it is also [+human] and third person. Interestingly, it is infelicitous if bound within the minimal clause when it is an argument of the verb. (That is, it seems to obey Principle B of the binding theory.) Although it is subject-oriented in the normal case, it can be bound by a non-subject if the verb is a “psych predicate,” that is, a predicate that denotes a feeling; in this case, it can be bound by the experiencer of the feeling. Again, in a discourse that depicts the thoughts, feelings, or point of view of a protagonist—the so-called “logophoric contexts”—it can be coreferential with the protagonist even if the latter is mentioned only in the preceding discourse (not within the sentence). These latter facts suggest that the anaphor is in fact coindexed with the perspective of the clause (rather than with the subject per se). In cases where this anaphor needs to be coindexed with the minimal subject (to express a meaning like ‘John loves himself’), the Dravidian languages exhibit two strategies to circumvent the Principle B effect. Malayalam adds an emphasis marker tanne to the anaphor; taan tanne can corefer with the minimal subject. This strategy parallels the strategy of European languages and East Asian languages (cf. Scandinavian seg selv). The three other major Dravidian languages—Tamil, Telugu, and Kannada—use a verbal reflexive: they add a light verb koL- (lit. ‘take’) to the verbal complex, which has the effect of reflexivizing the transitive predicate. (It either makes the verb intransitive or gives it a self-benefactive meaning.)
The Dravidian languages also have reciprocal and distributive anaphors. These have bipartite structures. An example of a Malayalam reciprocal anaphor is oral … matte aaL (‘one person … other person’). The distributive anaphor in Malayalam has the form awar-awar (‘they-they’); it is a reduplicated pronoun. The reciprocals and distributives are strict anaphors in the sense that they apparently obey Principle A; they must be bound in the domain of the minimal subject. They are not subject-oriented.
A noteworthy fact about the pronominal system of Dravidian is that the third person pronouns come in proximal-distal pairs, the proximal pronoun being used to refer to something nearby and the distal pronoun being used elsewhere.
Article
Toon Van Hal
The Ancient Greeks came into contact with possibilities and problems related to ‘language’ in several respects. The earliest epics contained implicit etymological explanations, and both the pre-Socratic philosophers and the sophists were intrigued by the link between the form of words and the meaning they carried. The adaptation of the Phoenician alphabet was an additional stimulus to start reflecting on language. ‘Letters’ became the smallest unit of inquiry in Greek language thought. Of the other units, the word was seen as the most significant level. Elaborating on the philosophical foundations laid by Plato, Aristotle, and early Stoic thinkers, Alexandrian scholars started shaping a philologically oriented tradition of grammar, which was largely oriented to the study of the eight parts of speech and directed at young students of Greek literature. Within the frame of grammar, less attention was paid to the level of the sentence, which explains why syntactic issues were not intensively explored. At its inception, Greek lexicography was an ancillary tool for understanding Greek literary texts too, directed at an audience of native speakers of Greek. Hence, lexicographical projects limited to including difficult or special words. Only once Romans began to delve into the study of Greek did the composition of general lexicons become more urgent.
Article
Susan Edwards and Christos Salis
Aphasia is an acquired language disorder subsequent to brain damage in the left hemisphere. It is characterized by diminished abilities to produce and understand both spoken and written language compared with the speaker’s presumed ability pre-cerebral damage. The type and severity of the aphasia depends not only on the location and extent of the cerebral damage but also the effect the lesion has on connecting areas of the brain. Type and severity of aphasia is diagnosed in comparison with assumed normal adult language. Language changes associated with normal aging are not classed as aphasia. The diagnosis and assessment of aphasia in children, which is unusual, takes account of age norms.
The most common cause of aphasia is a cerebral vascular
accident (CVA) commonly referred to as a stroke, but brain damage following traumatic head injury such as road accidents or gunshot wounds can also cause aphasia. Aphasia following such traumatic events is non-progressive in contrast to aphasia arising from brain tumor, some types of infection, or language disturbances in progressive conditions such as Alzheimer’s disease, where the language disturbance increases as the disease progresses.
The diagnosis of primary progressive aphasia (as opposed to non-progressive aphasia, the main focus of this article) is based on the following inclusion and exclusion criteria by M. Marsel Mesulam, in 2001. Inclusion criteria are as follows: Difficulty with language that interferes with activities of daily living and aphasia is the most prominent symptom. Exclusion criteria are as follows: Other non-degenerative disease or medical disorder, psychiatric diagnosis, episodic memory, visual memory, and visuo-perceptual impairment, and, finally, initial behavioral disturbance.
Aphasia involves one or more of the building blocks of language, phonemes, morphology, lexis, syntax, and semantics; and the deficits occur in various clusters or patterns across the spectrum. The degree of impairment varies across modalities, with written language often, but not always, more affected than spoken language. In some cases, understanding of language is relatively preserved, in others both production and understanding are affected. In addition to varied degrees of impairment in spoken and written language, any or more than one component of language can be affected. At the most severe end of the spectrum, a person with aphasia may be unable to communicate by either speech or writing and may be able to understand virtually nothing or only very limited social greetings. At the least severe end of the spectrum, the aphasic speaker may experience occasional word finding difficulties, often difficulties involving nouns; but unlike difficulties in recalling proper nouns in normal aging, word retrieval problems in mild aphasia includes other word classes.
Descriptions of different clusters of language deficits have led to the notion of syndromes. Despite great variations in the condition, patterns of language deficits associated with different areas of brain damage have been influential in understanding language-brain relationships. Increasing sophistication in language assessment and neurological investigations are contributing to a greater, yet still incomplete understanding of language-brain relationships.
Article
Patrik Bye
Morpheme ordering is largely explainable in terms of syntactic/semantic scope, or the Mirror Principle, although there is a significant residue of cases that resist an explanation in these terms. The article, we look at some key examples of (apparent) deviant ordering and review the main ways that linguists have attempted to account for them. Approaches to the phenomenon fall into two broad types. The first relies on mechanisms we can term “morphological,” while the second looks instead to the resources of the ‘narrow’ syntax or phonology. One morphological approach involves a template that associates each class of morphemes in the word with a particular position. A well-known example is the Bantu CARP (Causative-Applicative-Reciprocal-Passive) template, which requires particular orders between morphemes to obtain irrespective of scope. A second approach builds on the intuition that the boundary or join between a morpheme and the base to which it attaches can vary in closeness or strength, where ‘strength’ can be interpreted in gradient or discrete terms. Under the gradient interpretation, affixes differ in parsability, or separability from the base; understood discretely, as in Lexical Morphology and Phonology, morphemes (or classes of morphemes) may attach at a deeper morphological layer to stems (the stronger join), or to words (weaker join), which are closer to the surface. Deviant orderings may then arise where an affix attaches at a morphological layer deeper than its scope would lead us to expect. An example is the marking of case and possession in Finnish nouns: case takes scope over possession, but the case suffix precedes the possessive suffix. Another morphological approach is represented by Distributed Morphology, which permits certain local reorderings once all syntactic operations have taken place. Such operations may target specific morphemes, or morphosyntactic features characterizing a class of morphemes. Agreement marking is an interesting case, since agreement features are bundled as syntactically unitary heads but may in certain languages be split morphologically into separate affixes. This means that in the case of split agreement marking, the relative order must be attributed to post-syntactic principles. Besides these morphological approaches, other researchers have emphasized the resources of the narrow syntax, in particular phrasal movement, as a means for dealing with many challenging cases of morpheme ordering. Still other cases of apparently deviant ordering may be analyzed as epiphenomena of phonological processes and constraint interaction as they apply to prespecified and/or underspecified lexical representations.
Article
Hideki Kishimoto
Japanese is a language where the grammatical status of arguments and adjuncts is marked exclusively by postnominal case markers, and various argument realization patterns can be assessed by their case marking. Since Japanese is categorized as a language of the nominative-accusative type typologically, the unmarked case-marking frame obtained for transitive predicates of the non-stative (or eventive) type is ‘nominative-accusative’. Nevertheless, transitive predicates falling into the stative class often have other case-marking alignments, such as ‘nominative-nominative’ and ‘dative-nominative’. Consequently, Japanese provides much more varying argument realization patterns than those expected from its typological character as a nominative-accusative language.
In point of fact, argument marking can actually be much more elastic and variable, the variations being motivated by several linguistic factors. Arguments often have the option of receiving either syntactic or semantic case, with no difference in the logical or cognitive meaning (as in plural agent and source agent alternations) or depending on the meanings their predicate carry (as in locative alternation). The type of case marking that is not normally available in main clauses can sometimes be obtained in embedded contexts (i.e., in exceptional case marking and small-clause constructions). In complex predicates, including causative and indirect passive predicates, arguments are case-marked differently from their base clauses by virtue of suffixation, and their case patterns follow the mono-clausal case array, despite the fact that they have multi-clausal structures.
Various case marking options are also made available for arguments by grammatical operations. Some processes instantiate a change on the grammatical relations and case marking of arguments with no affixation or embedding. Japanese has the grammatical process of subjectivization, creating extra (non-thematic) major subjects, many of which are identified as instances of ‘possessor raising’ (or argument ascension). There is another type of grammatical process, which reduces the number of arguments by virtue of incorporating a noun into the predicate, as found in the light verb constructions with suru ‘do’ and the complex adjective constructions formed on the negative adjective nai ‘non-existent.’
Article
Malka Rappaport Hovav
Words are sensitive to syntactic context. Argument realization is the study of the relation between argument-taking words, the syntactic contexts they appear in and the interpretive properties that constrain the relation between them.
Article
Jim Wood and Neil Myler
The topic “argument structure and morphology” refers to the interaction between the number and nature of the arguments taken by a given predicate on the one hand, and the morphological makeup of that predicate on the other. This domain turns out to be crucial to the study of a number of theoretical issues, including the nature of thematic representations, the proper treatment of irregularity (both morphophonological and morphosemantic), and the very place of morphology in the architecture of the grammar. A recurring question within all existing theoretical approaches is whether word formation should be conceived of as split across two “places” in the grammar, or as taking place in only one.
Article
Philip Rubin
Arthur Seymour Abramson (1925–2017) was an American linguist who was prominent in the international experimental phonetics research community. He was best known for his pioneering work, with Leigh Lisker, on voice onset time (VOT), and for his many years spent studying tone and voice quality in languages such as Thai. Born and raised in Jersey City, New Jersey, Abramson served several years in the Army during World War II. Upon his return to civilian life, he attended Columbia University (BA, 1950; PhD, 1960). There he met Franklin Cooper, an adjunct who taught acoustic phonetics while also working for Haskins Laboratories. Abramson started working on a part-time basis at Haskins and remained affiliated with the institution until his death. For his doctoral dissertation (1962), he studied the vowels and tones of the Thai language, which would sit at the heart of his research and travels for the rest of his life. He would expand his investigations to include various languages and dialects, such as Pattani Malay and the Kuai dialect of Suai, a Mon-Khmer language. Abramson began his collaboration with University Pennsylvania linguist Leigh Lisker at Haskins Laboratories in the 1960s. Using their unique VOT technique, a sensitive measure of the articulatory timing between an occlusion in the vocal tract and the beginning of phonation (characterized by the onset of vibration of the vocal folds), they studied the voicing distinctions of various languages. Their long standing collaboration continued until Lisker’s death in 2006. Abramson and colleagues often made innovative use of state-of-art tools and technologies in their work, including transillumination of the larynx in running speech, X-ray movies of speakers in several languages/dialects, electroglottography, and articulatory speech synthesis.
Abramson’s career was also notable for the academic and scientific service roles that he assumed, including membership on the council of the International Phonetic Association (IPA), and as a coordinator of the effort to revise the International Phonetic Alphabet at the IPA’s 1989 Kiel Convention. He was also editor of the journal Language and Speech, and took on leadership roles at the Linguistic Society of America and the Acoustical Society of America. He was the founding Chair of the Linguistics Department at the University of Connecticut, which became a hotbed for research in experimental phonetics in the 1970s and 1980s because of its many affiliations with Haskins Laboratories. He also served for many years as a board member at Haskins, and Secretary of both the Board and the Haskins Corporation, where he was a friend and mentor to many.
Article
Marie K. Huffman
Articulatory phonetics is concerned with the physical mechanisms involved in producing spoken language. A fundamental goal of articulatory phonetics is to relate linguistic representations to articulator movements in real time and the consequent acoustic output that makes speech a medium for information transfer. Understanding the overall process requires an appreciation of the aerodynamic conditions necessary for sound production and the way that the various parts of the chest, neck, and head are used to produce speech. One descriptive goal of articulatory phonetics is the efficient and consistent description of the key articulatory properties that distinguish sounds used contrastively in language. There is fairly strong consensus in the field about the inventory of terms needed to achieve this goal. Despite this common, segmental, perspective, speech production is essentially dynamic in nature. Much remains to be learned about how the articulators are coordinated for production of individual sounds and how they are coordinated to produce sounds in sequence. Cutting across all of these issues is the broader question of which aspects of speech production are due to properties of the physical mechanism and which are the result of the nature of linguistic representations. A diversity of approaches is used to try to tease apart the physical and the linguistic contributions to the articulatory fabric of speech sounds in the world’s languages. A variety of instrumental techniques are currently available, and improvement in safe methods of tracking articulators in real time promises to soon bring major advances in our understanding of how speech is produced.
Article
Marianne Pouplier
One of the most fundamental problems in research on spoken language is to understand how the categorical, systemic knowledge that speakers have in the form of a phonological grammar maps onto the continuous, high-dimensional physical speech act that transmits the linguistic message. The invariant units of phonological analysis have no invariant analogue in the signal—any given phoneme can manifest itself in many possible variants, depending on context, speech rate, utterance position and the like, and the acoustic cues for a given phoneme are spread out over time across multiple linguistic units. Speakers and listeners are highly knowledgeable about the lawfully structured variation in the signal and they skillfully exploit articulatory and acoustic trading relations when speaking and perceiving. For the scientific description of spoken language understanding this association between abstract, discrete categories and continuous speech dynamics remains a formidable challenge. Articulatory Phonology and the associated Task Dynamic model present one particular proposal on how to step up to this challenge using the mathematics of dynamical systems with the central insight being that spoken language is fundamentally based on the production and perception of linguistically defined patterns of motion. In Articulatory Phonology, primitive units of phonological representation are called gestures. Gestures are defined based on linear second order differential equations, giving them inherent spatial and temporal specifications. Gestures control the vocal tract at a macroscopic level, harnessing the many degrees of freedom in the vocal tract into low-dimensional control units. Phonology, in this model, thus directly governs the spatial and temporal orchestration of vocal tract actions.
Article
Alan Reed Libert
Artificial languages—languages which have been consciously designed—have been created for more than 900 years, although the number of them has increased considerably in recent decades, and by the early 21st century the total figure probably was in the thousands. There have been several goals behind their creation; the traditional one (which applies to some of the best-known artificial languages, including Esperanto) is to make international communication easier. Some other well-known artificial languages, such as Klingon, have been designed in connection with works of fiction. Still others are simply personal projects.
A traditional way of classifying artificial languages involves the extent to which they make use of material from natural languages. Those artificial languages which are created mainly by taking material from one or more natural languages are called a posteriori languages (which again include well-known languages such as Esperanto), while those which do not use natural languages as sources are a priori languages (although many a posteriori languages have a limited amount of a priori material, and some a priori languages have a small number of a posteriori components). Between these two extremes are the mixed languages, which have large amounts of both a priori and a posteriori material. Artificial languages can also be classified typologically (as natural languages are) and by how and how much they have been used.
Many linguists seem to be biased against research on artificial languages, although some major linguists of the past have been interested in them.
Article
Anton Karl Ingason and Einar Freyr Sigurðsson
Attributive compounds are words that include two parts, a head and a non-head, both of which include lexical roots, and in which the non-head is interpreted as a modifier of the head. The nature of this modification is sometimes described in terms of a covert relationship R. The nature of R has been the subject of much discussion in the literature, including proposals that a finite and limited number of interpretive options are available for R, as well as analyses in which the interpretation of R is unrestricted and varies with context. The modification relationship between the parts of an attributive compound also contrasts with the interpretation of compounds in other ways because some non-heads in compounds saturate argument positions of the head, others are semantically conjoined with them, and some restrict their domain of interpretation.
Article
Gerrit Jan Dimmendaal
Nilo-Saharan, a phylum spread mainly across an area south of the Afro-Asiatic and north of the Niger-Congo phylum, was established as a genetic grouping by Greenberg. In his earlier, continent-wide classification of African languages in articles published between 1949 and 1954, Greenberg had proposed a Macro-Sudanic family (renamed Chari-Nile in subsequent studies), consisting of a Central Sudanic and an Eastern Sudanic branch plus two isolated members, Berta and Kunama. This family formed the core of the Nilo-Saharan phylum as postulated by Greenberg in his The Languages of Africa, where a number of groups were added which had been treated as isolated units in his earlier classificatory work: Songhay, Eastern Saharan (now called Saharan), Maban and Mimi, Nyangian (now called Kuliak or Rub), Temainian (Temeinian), Coman (Koman), and Gumuz.
Presenting an “encyclopaedic survey” of morphological structures for the more than 140 languages belonging to this phylum is impossible in such a brief study, also given the tremendous genetic distance between some of the major subgroups. Instead, typological variation in the morphological structure of these genetically-related languages will be central. In concrete terms this involves synchronic and diachronic observations on their formal properties (section 2), followed by an introduction to the nature of derivation, inflection, and compounding properties in Nilo-Saharan (section 3). This traditional compartmentalization has its limits because it misses out on the interaction with lexical structures and morphosyntactic properties in its extant members, as argued in section 4. As pointed out in section 5, language contact also must have played an important role in the geographical spreading of several of these typological properties.
Article
Lawrence D. Rosenblum
Research on visual and audiovisual speech information has profoundly influenced the fields of psycholinguistics, perception psychology, and cognitive neuroscience. Visual speech findings have provided some of most the important human demonstrations of our new conception of the perceptual brain as being supremely multimodal. This “multisensory revolution” has seen a tremendous growth in research on how the senses integrate, cross-facilitate, and share their experience with one another.
The ubiquity and apparent automaticity of multisensory speech has led many theorists to propose that the speech brain is agnostic with regard to sense modality: it might not know or care from which modality speech information comes. Instead, the speech function may act to extract supramodal informational patterns that are common in form across energy streams. Alternatively, other theorists have argued that any common information existent across the modalities is minimal and rudimentary, so that multisensory perception largely depends on the observer’s associative experience between the streams. From this perspective, the auditory stream is typically considered primary for the speech brain, with visual speech simply appended to its processing. If the utility of multisensory speech is a consequence of a supramodal informational coherence, then cross-sensory “integration” may be primarily a consequence of the informational input itself. If true, then one would expect to see evidence for integration occurring early in the perceptual process, as well in a largely complete and automatic/impenetrable manner. Alternatively, if multisensory speech perception is based on associative experience between the modal streams, then no constraints on how completely or automatically the senses integrate are dictated. There is behavioral and neurophysiological research supporting both perspectives.
Much of this research is based on testing the well-known McGurk effect, in which audiovisual speech information is thought to integrate to the extent that visual information can affect what listeners report hearing. However, there is now good reason to believe that the McGurk effect is not a valid test of multisensory integration. For example, there are clear cases in which responses indicate that the effect fails, while other measures suggest that integration is actually occurring. By mistakenly conflating the McGurk effect with speech integration itself, interpretations of the completeness and automaticity of multisensory may be incorrect. Future research should use more sensitive behavioral and neurophysiological measures of cross-modal influence to examine these issues.
Article
William R. Leben
Autosegments were introduced by John Goldsmith in his 1976 M.I.T. dissertation to represent tone and other suprasegmental phenomena. Goldsmith’s intuition, embodied in the term he created, was that autosegments constituted an independent, conceptually equal tier of phonological representation, with both tiers realized simultaneously like the separate voices in a musical score.
The analysis of suprasegmentals came late to generative phonology, even though it had been tackled in American structuralism with the long components of Harris’s 1944 article, “Simultaneous components in phonology” and despite being a particular focus of Firthian prosodic analysis. The standard version of generative phonology of the era (Chomsky and Halle’s The Sound Pattern of English) made no special provision for phenomena that had been labeled suprasegmental or prosodic by earlier traditions.
An early sign that tones required a separate tier of representation was the phenomenon of tonal stability. In many tone languages, when vowels are lost historically or synchronically, their tones remain. The behavior of contour tones in many languages also falls into place when the contours are broken down into sequences of level tones on an independent level or representation. The autosegmental framework captured this naturally, since a sequence of elements on one tier can be connected to a single element on another. But the single most compelling aspect of the early autosegmental model was a natural account of tone spreading, a very common process that was only awkwardly captured by rules of whatever sort. Goldsmith’s autosegmental solution was the Well-Formedness Condition, requiring, among other things, that every tone on the tonal tier be associated with some segment on the segmental tier, and vice versa. Tones thus spread more or less automatically to segments lacking them. The Well-Formedness Condition, at the very core of the autosegmental framework, was a rare constraint, posited nearly two decades before Optimality Theory.
One-to-many associations and spreading onto adjacent elements are characteristic of tone but not confined to it. Similar behaviors are widespread in long-distance phenomena, including intonation, vowel harmony, and nasal prosodies, as well as more locally with partial or full assimilation across adjacent segments.
The early autosegmental notion of tiers of representation that were distinct but conceptually equal soon gave way to a model with one basic tier connected to tiers for particular kinds of articulation, including tone and intonation, nasality, vowel features, and others. This has led to hierarchical representations of phonological features in current models of feature geometry, replacing the unordered distinctive feature matrices of early generative phonology. Autosegmental representations and processes also provide a means of representing non-concatenative morphology, notably the complex interweaving of roots and patterns in Semitic languages.
Later work modified many of the key properties of the autosegmental model. Optimality Theory has led to a radical rethinking of autosegmental mapping, delinking, and spreading as they were formulated under the earlier derivational paradigm.
Article
Adina Dragomirescu
Balkan-Romance is represented by Romanian and its historical dialects: Daco-Romanian (broadly known as Romanian), Aromanian, Megleno-Romanian, and Istro-Romanian (see article “Morphological and Syntactic Variation and Change in Romanian” in this encyclopedia). The external history of these varieties is often unclear, given the historical events that took place in the Lower Danubian region: the conquest of this territory by the Roman Empire for a short period and the successive Slavic invasions. Moreover, the earliest preserved writing in Romanian only dates from the 16th century. Between the Roman presence in the Balkans and the first attested text, there is a gap of more than 1,000 years, a period in which Romanian emerged, the dialectal separation took place, and the Slavic influence had effects especially on the lexis of Romanian.
In the 16th century, in the earliest old Romanian texts, the language already displayed the main features of modern Romanian: the vowels /ə/ and /ɨ/; the nominative-accusative versus genitive-dative case distinction; analytical case markers, such as the genitive marker al; the functional prepositions a and la; the proclitic genitive-dative marker lui; the suffixal definite article; polydefinite structures; possessive affixes; rich verbal inflection, with both analytic and synthetic forms and with three auxiliaries (‘have’, ‘be’, and ‘want’); the supine, not completely verbalized at the time; two types of infinitives, with the ‘short’ one on a path toward becoming verbal and the ‘long’ one specializing as a noun; null subjects; nonfinite verb forms with lexical subjects; the mechanism for differential object marking and clitic doubling with slightly more vacillating rules than in the present-day language; two types of passives; strict negative concord; the SVO and VSO word orders; adjectives placed mainly in the postnominal position; a rich system of pronominal clitics; prepositions requiring the accusative and the genitive; and a large inventory of subordinating conjunctions introducing complement clauses.
Most of these features are also attested in the trans-Danubian varieties (Aromanian, Megleno-Romanian, and Istro-Romanian), which were also strongly influenced by the various languages they have entered in direct contact with: Greek, Albanian, Macedonian, Croatian, and so forth. These source languages have had a major influence in the vocabulary of the trans-Danubian varieties and certain consequences in the shape of their grammatical system. The differences between Daco-Romanian and the trans-Danubian varieties have also resulted from the preservation of archaic features in the latter or from innovations that took place only there.
Article
Bert Le Bruyn, Henriëtte de Swart, and Joost Zwarts
Bare nominals (also called “bare nouns”) are nominal structures without an overt article or other determiner. The distinction between a bare noun and a noun that is part of a larger nominal structure must be made in context: Milk is a bare nominal in I bought milk, but not in I bought the milk. Bare nouns have a limited distribution: In subject or object position, English allows bare mass nouns and bare plurals, but not bare singular count nouns (*I bought table). Bare singular count nouns only appear in special configurations, such as coordination (I bought table and chairs for £182).
From a semantic perspective, it is noteworthy that bare nouns achieve reference without the support of a determiner. A full noun phrase like the cookies refers to the maximal sum of cookies in the context, because of the definite article the. English bare plurals have two main interpretations: In generic sentences they refer to the kind (Cookies are sweet), in episodic sentences they refer to some exemplars of the kind (Cookies are in the cabinet). Bare nouns typically take narrow scope with respect to other scope-bearing operators like negation.
The typology of bare nouns reveals substantial variation, and bare nouns in languages other than English may have different distributions and meanings. But genericity and narrow scope are recurring features in the cross-linguistic study of bare nominals.
Article
Maarten Kossmann
Since the start of the Islamic conquest of the Maghreb in the 7th century ce, Berber and Arabic have been in continual contact. This has led to large-scale mutual influence. The sociolinguistic setting of this influence is not the same, though; Arabic influence on Berber is found in a situation of language maintenance with widespread bilingualism, while Berber influence on Arabic is no doubt to a large degree due to language shift by Berber speakers to Arabic.
Linguistic influence is found on all levels: phonology, morphology, syntax, and lexicon. In those cases where only innovative patterns are shared between the two language groups, it is often difficult to make out where the innovation started; thus the great similarities in syllable structure between Maghrebian Arabic and northern Berber are the result of innovations within both language families, and it is difficult to tell where it started. Morphological influence seems to be mediated exclusively by lexical borrowing. Especially in Berber, this has led to parallel systems in the morphology, where native words always have native morphology, while loans either have nativized morphology or retain Arabic-like patterns. In the lexicon, it is especially Berber that takes over scores of loanwords from Arabic, amounting in one case to over one-third of the basic lexicon as defined by 100-word lists.
Article
Bilingualism/multilingualism is a natural phenomenon worldwide. Unwittingly, however, monolingualism has been used as a standard to characterize and define bilingualism/multilingualism in linguistic research. Such a conception led to a “fractional,” “irregular,” and “distorted” view of bilingualism, which is becoming rapidly outmoded in the light of multipronged, rapidly growing interdisciplinary research. This article presents a complex and holistic view of bilinguals and multilinguals on conceptual, theoretical, and pragmatic/applied grounds. In that process, it attempts to explain why bilinguals are not a mere composite of two monolinguals. If bilinguals were a clone of two monolinguals, the study of bilingualism would not merit any substantive consideration in order to come to grips with bilingualism; all one would have to do is focus on the study of a monolingual person. Interestingly, even the two bilinguals are not clones of each other, let alone bilinguals as a set of two monolinguals. This paper examines the multiple worlds of bilinguals in terms of their social life and social interaction. The intricate problem of defining and describing bilinguals is addressed; their process and end result of becoming bilinguals is explored alongside their verbal interactions and language organization in the brain. The role of social and political bilingualism is also explored as it interacts with individual bilingualism and global bilingualism (e.g., the issue of language endangerment and language death).
Other central concepts such as individuals’ bilingual language attitudes, language choices, and consequences are addressed, which set bilinguals apart from monolinguals. Language acquisition is as much an innate, biological, as social phenomenon; these two complementary dimensions receive consideration in this article along with the educational issues of school performance by bilinguals. Is bilingualism a blessing or a curse? The linguistic and cognitive consequences of individual, societal, and political bilingualism are examined.