Generative Grammar
A Generative Grammar is a formal grammar that will correctly predict which combinations of words will form grammatical sentences.
- Example(s):
- See: Morphology (Linguistics), Transformational Grammar, Linguistics, Minimalist Program, Dependency Grammar, Recursive Function.
References
2014
- (Wikipedia, 2014) ⇒ http://en.wikipedia.org/wiki/generative_grammar Retrieved:2014-7-27.
- In theoretical linguistics, a generative grammar refers to a particular approach to the study of syntax. A generative grammar of a language attempts to give a set of rules that will correctly predict which combinations of words will form grammatical sentences. In most approaches to generative grammar, the rules will also predict the morphology of a sentence. Generative grammar arguably originates in the work of Noam Chomsky, beginning in the late 1950s. However, Chomsky has said that the first generative grammar in the modern sense was Panini's Sanskrit grammar.[1] Chomsky also acknowledges other historical antecedents. [2]
Early versions of Chomsky's theory were called transformational grammar, and this term is still used as a general term that includes his subsequent theories. There are a number of competing versions of generative grammar currently practiced within linguistics. Chomsky's current theory is known as the Minimalist program. Other prominent theories include or have included dependency grammar, head-driven phrase structure grammar, lexical functional grammar, categorial grammar, relational grammar, link grammar, and tree-adjoining grammar. Chomsky has argued that many of the properties of a generative grammar arise from an "innate" universal grammar. Proponents of generative grammar have argued that most grammar is not the result of communicative function and is not simply learned from the environment (see the poverty of the stimulus argument). In this respect, generative grammar takes a point of view different from cognitive grammar, functional, and behaviorist theories. Most versions of generative grammar characterize sentences as either grammatically correct (also known as well formed) or not. The rules of a generative grammar typically function as an algorithm to predict grammaticality as a discrete (yes-or-no) result. In this respect, it differs from stochastic grammar, which considers grammaticality as a probabilistic variable. However, some work in generative grammar (e.g. recent work by Joan Bresnan) uses stochastic versions of optimality theory.
- In theoretical linguistics, a generative grammar refers to a particular approach to the study of syntax. A generative grammar of a language attempts to give a set of rules that will correctly predict which combinations of words will form grammatical sentences. In most approaches to generative grammar, the rules will also predict the morphology of a sentence. Generative grammar arguably originates in the work of Noam Chomsky, beginning in the late 1950s. However, Chomsky has said that the first generative grammar in the modern sense was Panini's Sanskrit grammar.[1] Chomsky also acknowledges other historical antecedents. [2]
- ↑ S.S. Chattopadhyay, An event in Kolkata, Frontline
- ↑ Another example is Humboldt. Chomsky quotes Humboldt's description of language as a system which "makes infinite use of finite means".
1972
Jackendoff, Ray. Semantic interpretation in generative grammar. Cambridge, MA: MIT press, 1972.
1957
- (Chomsky, 1957) ⇒ Noam Chomsky. (1957). “Syntactic Structures.