Universal grammar
Universal grammar (UG), in modern linguistics, is the theory of the innate biological component of the language faculty, usually credited to Noam Chomsky. The basic postulate of UG is that there are innate constraints on what the grammar of a possible human language could be. When linguistic stimuli are received in the course of language acquisition, children then adopt specific syntactic rules that conform to UG. The advocates of this theory emphasize and partially rely on the poverty of the stimulus (POS) argument and the existence of some universal properties of natural human languages. However, the latter has not been firmly established.
Other linguists have opposed that notion, arguing that languages are so diverse that the postulated universality is rare. The theory of universal grammar remains a subject of debate among linguists.
Overview
The term "universal grammar" is placeholder for whichever domain-specific features of linguistic competence turn out to be innate. Within generative grammar, it is generally accepted that there must be some such features, and one of the goals of generative research is to formulate and test hypotheses about which aspects those are. In day-to-day generative research, the notion that universal grammar exists motivates analyses in terms of general principles. As much as possible, facts about particular languages are derived from these general principles rather than from language-specific stipulations.
Evidence
The idea that at least some aspects are innate is motivated by poverty of the stimulus arguments. For example, one famous poverty of the stimulus argument concerns the acquisition of yes–no questions in English. This argument starts from the observation that children only make mistakes compatible with rules targeting hierarchical structure even though the examples which they encounter could have been generated by a simpler rule that targets linear order. In other words, children seem to ignore the possibility that the question rule is as simple as "switch the order of the first two words" and immediately jump to alternatives that rearrange constituents in tree structures. This is taken as evidence that children are born knowing that grammatical rules involve hierarchical structure, even though they have to figure out what those rules are.
History
Between 1100 and 1400, the theoretical work on matters of language significantly expanded in western Europe, its typical social context being the teaching of grammar, logic, or theology, producing a vast literature on aspects of linguistic theory, such as a 13th century theory of grammar known in modern times as modism, although no assertions were made in the texts about "a theory of language," as such. While not much work was done on the evolution of languages, Dante and Roger Bacon offered perceptive observations. Bacon had a complex notion of grammar, which ranged from the teaching of elementary Latin through what he termed "rational grammar," to research on the so-called languages of sacred wisdom, i.e. Latin and Greek. Professor of Latin literature Raf Van Rooy quotes Bacon's "notorious" dictum on grammar used to denote regional linguistic variation and notes Bacon's contention that Latin and Greek, although "one in substance," were each characterized by many idioms (idiomata: proprietates). Van Rooy speculates that Bacon's references to grammar concerned a "quasi-universal nature of grammatical categories," whereas his assertions on Greek and Latin were applications of his lingua/idioma distinction rather than a generalizing statements on the nature of grammar. Linguistics professor Margaret Thomas acknowledges that "intellectual commerce between ideas about Universal Grammar and [second language] acquisition is not a late-20th century invention," but rejects as "convenient" the interpretation of Bacon's dictum by generative grammarians as an assertion by the English polymath of the existence of universal grammar.
The concept of a generalized grammar was at the core of the 17th century projects for philosophical languages. An influential work in that time was Grammaire générale by Claude Lancelot and Antoine Arnauld. They describe a general grammar for languages, coming to the conclusion that grammar has to be universal. There is a Scottish school of universal grammarians from the 18th century that included James Beattie, Hugh Blair, James Burnett, James Harris, and Adam Smith, distinguished from the philosophical-language project.
The article on grammar in the first edition of the Encyclopædia Britannica (1771) contains an extensive section titled "Of Universal Grammar."
In the late 19th and early 20th century, Wilhelm Wundt and Otto Jespersen claimed that these earlier arguments were overly influenced by Latin and ignored the breadth of worldwide language real grammar", but reduced it to universal syntactic categories or super-categories, such as number, tenses, etc.
Behaviorists, after the rise of the eponymous theory, advanced the idea that language acquisition, like any other kind of learning, could be explained by a succession of trials, errors, and rewards for success. In other words, children, according to behaviorists, learn their mother tongue by simple imitation, i.e. through listening and repeating what adults say. For example, when a child says "milk" and the mother will smile and give milk to her child, then, as a result, the child will find this outcome rewarding, which enhances the child's language development.
Theories of universal grammar
Within generative grammar, there are a variety of theories about what universal grammar consists of. One notable hypothesis proposed by Hagit Borer holds that the fundamental syntactic operations are universal and that all variation arises from different feature-specifications in the lexicon. On the other hand, a strong hypothesis adopted in some variants of Optimality Theory holds that humans are born with a universal set of constraints, and that all variation arises from differences in how these constraints are ranked. In a 2002 paper, Noam Chomsky, Marc Hauser and W. Tecumseh Fitch proposed that universal grammar consists solely of the capacity for hierarchical phrase structure.
The main hypotheses
In an article entitled "The Faculty of Language: What Is It, Who Has It, and How Did It Evolve?" Hauser, Chomsky, and Fitch present the three leading hypotheses for how language evolved and brought humans to the point where they have a universal grammar.
The first hypothesis states that the faculty of language in the broad sense (FLb) is strictly homologous to animal communication. This means that homologous aspects of the faculty of language exist in non-human animals.
The second hypothesis states that the FLb is a derived and uniquely human adaptation for language. This hypothesis holds that individual traits were subject to natural selection and came to be specialized for humans.
The third hypothesis states that only the faculty of language in the narrow sense (FLn) is unique to humans. It holds that while mechanisms of the FLb are present in both human and non-human animals, the computational mechanism of recursion has evolved recently, and solely in humans.
Presence of creole languages
The presence of creole languages is sometimes cited as further support for this theory, especially by Bickerton's language bioprogram theory. Creole languages develop and form when disparate societies with no common language come together and are forced to devise a new system of communication. The system used by the original speakers is typically an inconsistent mix of vocabulary items, known as pidgin. As these speakers' children begin to acquire their first language, they use the pidgin input to effectively create their own original language, known as a creole language. Unlike pidgins, creole languages have native speakers (those with language acquisition from early childhood) and make use of a full, systematic grammar.
Bickerton claims the fact that certain features are shared by virtually all creole languages supports the notion of a universal grammar. For example, their default point of reference in time (expressed by bare verb stems) is not the present moment, but the past. Using pre-verbal auxiliaries, they uniformly express tense, aspect, and mood. Negative concord occurs, but it affects the verbal subject (as opposed to the object, as it does in languages like Spanish). Another similarity among creole languages can be identified in the fact that questions are created simply by changing the intonation of a declarative sentence; not its word order or content.
Opposing this notion, the work by Carla Hudson-Kam and Elissa Newport suggests that creole languages may not support a universal grammar at all. In a series of experiments, Hudson-Kam and Newport looked at how children and adults learn artificial grammars. They found that children tend to ignore minor variations in the input when those variations are infrequent, and reproduce only the most frequent forms. In doing so, the children tend to standardize the language they hear around them. Hudson-Kam and Newport hypothesize that in a pidgin-development situation (and in the real-life situation of a deaf child whose parents are or were disfluent signers), children systematize the language they hear, based on the probability and frequency of forms, and not that which has been suggested on the basis of a universal grammar. Further, they argue, it seems to follow that creole languages would share features with the languages from which they are derived, and thus look similar in terms of grammar.
Many researchers of universal grammar argue against the concept of relexification, i.e. that a language replaces its lexicon almost entirely with that of another. This, they argue, goes against the universalist notions of a universal grammar, which has an innate grammar.
Views and assessments
Recent research has used recurrent neural network architectures (RNNs). McCoy et al. (2018) focused on a strong version of the poverty-of-the-stimulus argument, which claims that language learners require a hierarchical constraint, although they report that a milder version, which only asserts that a hierarchical bias is necessary, is difficult to assess using RNNs because RNNs must possess some biases and the nature of these biases remains "currently poorly understood." They go on to acknowledge that while all the architectures they used had a bias toward linear order and the GRU-with-attention architecture was the only one that overcame this linear bias sufficiently to generalize hierarchically. "Humans certainly could have such an innate constraint."
The empirical basis of poverty-of-the-stimulus arguments has been challenged by Geoffrey Pullum and others, leading to a persistent back-and-forth debate in the language acquisition literature.
Language acquisition researcher Michael Ramscar has suggested that when children erroneously expect an ungrammatical form that then never occurs, the repeated failure of expectation serves as a form of implicit negative feedback that allows them to correct their errors over time, in the way that, for example, children correct grammar generalizations like goed to went through repetitive failure.
In addition, it has been suggested that people learn about probabilistic patterns of word distribution in their language, rather than hard and fast rules (see Distributional hypothesis). For example, in English, children overgeneralize the past tense marker "-ed" and conjugate irregular verbs as if they were regular, producing forms like goed and eated, and then correct this deviancy over time. It has also been hypothesized that the poverty of the stimulus problem can be largely avoided if it is assumed that children employ similarity-based generalization strategies in language learning, i.e. generalizing about the usage of new words from similar words they already know how to use.
Neurogeneticists Simon Fisher and Sonja Vernes observe that, with human language-skills being evidently unmatched elsewhere in the world's fauna, there have been several theories about one single mutation event occurring some time in the past in our nonspeaking ancestors, as argued by e.g. Chomsky (2011), i.e. some "lone spark that was sufficient to trigger the sudden appearance of language and culture." They characterize that notion as "romantic" and "inconsistent with the messy mappings between genetics and cognitive processes." According to Fisher & Vernes, the link between genes to grammar has not been consistently mapped by scientists. What has been established by research, they claim, relates primarily to speech pathologies. The arising lack of certainty, they conclude, has provided an audience for "unconstrained speculations" that have fed the "myth" of "so-called grammar genes".
Professor of Natural Language Computing Geoffrey Sampson maintains that universal grammar theories are not falsifiable and are therefore pseudoscientific. He argues that the grammatical "rules" linguists posit are simply post-hoc observations about existing languages, rather than predictions about what is possible in a language. Sampson claims that every one of the "poor" arguments used to justify the language-instinct claim is wrong. He writes that "either the logic is fallacious, or the factual data are incorrect (or, sometimes, both)," and the "evidence points the other way." Children are good at learning languages, because people are good at learning anything that life throws at us — not because we have fixed structures of knowledge built-in. Similarly, professor of cognitive science Jeffrey Elman argues that the unlearnability of languages ostensibly assumed by universal grammar is based on a too-strict, "worst-case" model of grammar, which is not in keeping with any actual grammar.
Linguist James Hurford, in his article "Nativist and Functional Explanations in Language Acquisition," offers the major differences between the glossogenetic and the phylogenetic mechanisms. He states that, "Deep aspects of the form of language are not likely to be readily identifiable with obvious specific uses, and one cannot suppose that it will be possible to attribute them directly to the recurring short-term needs of successive generations in a community. Here, nativist explanations for aspects of the form of language, appealing to an innate LAD, seem appropriate. But use or function can also be appealed to on the evolutionary timescale, to attempt to explain the structure of the LAD itself." For Hurford, biological mutations plus functional considerations constitute the explanans, while the LAD itself constitutes the explanandum. The LAD is part of the species' heredity, the result of mutations over a long period, he states. But, while he agrees with Chomsky that the mechanism of grammaticisation is located in "the Chomskyan LAD" and that Chomsly is "entirely right in emphasising that a language (E-language) is an artifact resulting from the interplay of many factors," he states that this artifact should be of great interest and systematic study, and can affect grammatical competence, i.e. "I-language."
Morten H. Christiansen, professor of Psychology, and Nick Chater, professor of Psychology and Language Sciences, have argued that "a biologically determined UG is not evolutionarily viable." As the processes of language change are much more rapid than processes of genetic change, they state, language constitutes a "moving target" both over time and across different human populations, and, hence, cannot provide a stable environment to which language genes could have adapted. In following Darwin, they view language as a complex and interdependent "organism," which evolves under selectional pressures from human learning and processing mechanisms, so that "apparently arbitrary aspects of linguistic structure may result from general learning and processing biases deriving from the structure of thought processes, perceptuo-motor factors, cognitive limitations, and pragmatics". Professor of linguistics Norbert Hornstein countered polemically that Christiansen and Chater appear "to have no idea what generative grammar [theory] is," and "especially, but not uniquely, about the Chomsky program." Hornstein points out that all "grammatically informed psycho-linguistic works done today or before" understand that generative/universal grammar capacities are but one factor among others needed to explain real-time acquisition." Christiansen and Chater's observation that "language use involves multiple interacting variables" [italics in the original] is, essentially, a truism. It is nothing new, he argues, to state that "much more than a competence theory will be required" to figure out how language is deployed, acquired, produced, or parsed. The position, he concludes, that universal grammar properties are just "probabilistic generalizations over available linguistic inputs" belongs to the "traditional" and "debunked" view held by associationists and structuralists many decades in the past.
In the same vein, professor of linguistics Nicholas Evans and prfessor of psycholinguistics Stephen C. Levinson observe that Chomsky’s notion of a Universal Grammar has been mistaken for a set of substantial research findings about what all languages have in common, while, it is, "in fact," the programmatic label for "whatever it turns out to be that all children bring to learning a language." For substantial findings about universals across languages, they argue, one must turn to the field of linguistic typology, which bares a "bewildering range of diverse languages" and in which "generalizations are really quite hard to extract." Chomsky’s actual views, combining, as they claim, philosophical and mathematical approaches to structure with claims about the innate endowment for language, have been "hugely influential in the cognitive sciences.
Wolfram Hinzen, in his work The philosophical significance of Universal Grammar seeks to re-establish the epistemological significance of grammar and addresses the three main current objections to Cartesian universal grammar, i.e. that it has no coherent formulation, it cannot have evolved by standard, accepted neo-Darwinian evolutionary principles, and it goes against the variation extant at all levels of linguistic organization, which lies at the heart of human faculty of language.
In the domain of field research, Daniel Everett has claimed that the Pirahã language is a counterexample to the basic tenets of universal grammar because it lacks clausal embedding. According to Everett, this trait results from Pirahã culture emphasizing present-moment concrete matters. Nevins et al. (2007) have responded that Pirahã does, in fact, have clausal embedding, and that, even if it did not, this would be irrelevant to current theories of universal grammar. They addressed each of Everett's claims and, using Everett's "rich material" data, claim to have found no evidence of a causal relation between culture and grammatical structure. Pirahã grammar, they concluded, presents no unusual challenge, much less the "severe" one claimed by Everett, to the notion of a universal grammar.
Developments
In 2017, Chomsky and Berwick co-wrote their book titled Why Only Us, where they defined both the minimalist program and the strong minimalist thesis and its implications, to update their approach to UG theory. According to Berwick and Chomsky, "the optimal situation would be that UG reduces to the simplest computational principles which operate in accord with conditions of computational efficiency. This conjecture is ... called the Strong Minimalist Thesis (SMT)."
The significance of SMT is to shift the previous emphasis on a universal grammar to the concept that Chomsky and Berwick now call "merge". "Merge" is defined there as follows:
SMT dictates that "Merge will be as simple as possible: it will not modify X or Y or impose any arrangement on them; in particular, it will leave them unordered; an important fact. Merge is therefore just set formation: Merge of X and Y yields the set {X, Y}."
See also
Notes
References
Bibliography
Further reading
- Moro, Andrea (2016). Impossible Languages. The MIT Press. ISBN 978-0262034890.