Conservativity
In formal semantics conservativity is a proposed linguistic universal which states that any determiner D must obey the equivalence D(A,B) \leftrightarrow D(A, A\cap B). For instance, the English determiner "every" can be seen to be conservative by the equivalence of the following two sentences, schematized in generalized quantifier notation to the right. # Every aardvark bites. \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \rightsquigarrow every(A,B) # Every aardvark is an aardvark that bites. \ \ \rightsquigarrow every(A,A\cap B) Conceptually, conservativity can be understood as saying that the elements of B which are not elements of A are not relevant for evaluating the truth of the determiner phrase as a whole. For instance, truth of the first sentence above does not depend on which biting non-aardvarks exist. Conservativity is significant to semantic theory because there are many logically possible determiners which are not attested as denotations of natur ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Generalized Quantifier
In formal semantics, a generalized quantifier (GQ) is an expression that denotes a set of sets. This is the standard semantics assigned to quantified noun phrases. For example, the generalized quantifier ''every boy'' denotes the set of sets of which every boy is a member: \ This treatment of quantifiers has been essential in achieving a compositional semantics for sentences containing quantifiers. Type theory A version of type theory is often used to make the semantics of different kinds of expressions explicit. The standard construction defines the set of types recursively as follows: #''e'' and ''t'' are types. #If ''a'' and ''b'' are both types, then so is \langle a,b\rangle #Nothing is a type, except what can be constructed on the basis of lines 1 and 2 above. Given this definition, we have the simple types ''e'' and ''t'', but also a countable infinity of complex types, some of which include: \langle e,t\rangle;\qquad \langle t,t\rangle;\qquad \langle\langle e,t\rangle, t ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Determiner
A determiner, also called determinative (abbreviated ), is a word, phrase, or affix that occurs together with a noun or noun phrase and generally serves to express the reference of that noun or noun phrase in the context. That is, a determiner may indicate whether the noun is referring to a definite or indefinite element of a class, to a closer or more distant element, to an element belonging to a specified person or thing, to a particular number or quantity, etc. Common kinds of determiners include definite and indefinite articles (''the'', ''a''), demonstratives (''this'', ''that''), possessive determiners (''my,'' ''their''), cardinal numerals (''one'', ''two''), quantifiers (''many'', ''both''), distributive determiners (''each'', ''every''), and interrogative determiners (''which'', ''what''). Description Most determiners have been traditionally classed either as adjectives or pronouns, and this still occurs in traditional grammars: for example, demonstrative and possessi ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Determiner Phrase
In linguistics, a determiner phrase (DP) is a type of phrase headed by a determiner such as ''many''. Controversially, many approaches, take a phrase like ''not very many apples'' to be a DP, headed, in this case, by the determiner ''many''. This is called the DP analysis or the DP hypothesis. Others reject this analysis in favor of the more traditional NP (noun phrase or nominal phrase) analysis where ''apples'' would be the head of the phrase in which the DP ''not very many'' is merely a dependent. Thus, there are competing analyses concerning heads and dependents in nominal groups. The DP analysis developed in the late 1970s and early 1980s, and it is the majority view in generative grammar today. In the example determiner phrases below, the determiners are in boldface: * a little dog, the little dogs (indefinite or definite articles) * my little dog, your little dogs (possessives) * this little dog, those little dogs (demonstratives) * every little dog, each little dog, no dog ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Linguistic Universal
A linguistic universal is a pattern that occurs systematically across natural languages, potentially true for all of them. For example, ''All languages have nouns and verbs'', or ''If a language is spoken, it has consonants and vowels.'' Research in this area of linguistics is closely tied to the study of linguistic typology, and intends to reveal generalizations across languages, likely tied to cognition, perception, or other abilities of the mind. The field originates from discussions influenced by Noam Chomsky's proposal of a Universal Grammar, but was largely pioneered by the linguist Joseph Greenberg, who derived a set of forty-five basic universals, mostly dealing with syntax, from a study of some thirty languages. Though there has been significant research into linguistic universals, in more recent time some linguists, including Nicolas Evans and Stephen C. Levinson, have argued against the existence of absolute linguistic universals that are shared across all languages. ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Formal Semantics (natural Language)
Formal semantics is the study of grammatical meaning in natural languages using formal tools from logic and theoretical computer science. It is an interdisciplinary field, sometimes regarded as a subfield of both linguistics and philosophy of language. It provides accounts of what linguistic expressions mean and how their meanings are composed from the meanings of their parts. The enterprise of formal semantics can be thought of as that of reverse-engineering the semantic components of natural languages' grammars. Overview Formal semantics studies the denotations of natural language expressions. High-level concerns include compositionality, reference, and the nature of meaning. Key topic areas include scope, modality, binding, tense, and aspect. Semantics is distinct from pragmatics, which encompasses aspects of meaning which arise from interaction and communicative intent. Formal semantics is an interdisciplinary field, often viewed as a subfield of both linguistics and ph ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Universal Grammar
Universal grammar (UG), in modern linguistics, is the theory of the genetic component of the language faculty, usually credited to Noam Chomsky. The basic postulate of UG is that there are innate constraints on what the grammar of a possible human language could be. When linguistic stimuli are received in the course of language acquisition, children then adopt specific syntactic rules that conform to UG. The advocates of this theory emphasize and partially rely on the poverty of the stimulus (POS) argument and the existence of some universal properties of natural human languages. However, the latter has not been firmly established, as some linguists have argued languages are so diverse that such universality is rare. It is a matter of empirical investigation to determine precisely what properties are universal and what linguistic capacities are innate. Argument The theory of universal grammar proposes that if human beings are brought up under normal conditions (not those of ex ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Lindström Quantifier
In mathematical logic, a Lindström quantifier is a generalized polyadic quantifier. Lindström quantifiers generalize first-order quantifiers, such as the existential quantifier, the universal quantifier, and the counting quantifiers. They were introduced by Per Lindström in 1966. They were later studied for their applications in logic in computer science and database query languages. Generalization of first-order quantifiers In order to facilitate discussion, some notational conventions need explaining. The expression : \phi^=\ for ''A'' an ''L''-structure (or ''L''-model) in a language ''L'', ''φ'' an ''L''-formula, and \bar a tuple of elements of the domain dom(''A'') of ''A''. In other words, \phi^ denotes a ( monadic) property defined on dom(A). In general, where ''x'' is replaced by an ''n''-tuple \bar of free variables, \phi^ denotes an ''n''-ary relation defined on dom(''A''). Each quantifier Q_A is relativized to a structure, since each quantifier is viewed a ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Jon Barwise
Kenneth Jon Barwise (; June 29, 1942 – March 5, 2000) was an American mathematician, philosopher and logician who proposed some fundamental revisions to the way that logic is understood and used. Education and career Born in Independence, Missouri to Kenneth T. and Evelyn Barwise, Jon was a precocious child. A pupil of Solomon Feferman at Stanford University, Barwise started his research in infinitary logic. After positions as assistant professor at Yale University and the University of Wisconsin, during which time his interests turned to natural language, he returned to Stanford in 1983 to direct the Center for the Study of Language and Information. He began teaching at Indiana University in 1990. He was elected a Fellow of the American Academy of Arts and Sciences in 1999. In his last year, Barwise was invited to give the 2000 Gödel Lecture; he died prior to the lecture. Philosophical and logical work Barwise contended that, by being explicit about the context in whic ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Language Acquisition
Language acquisition is the process by which humans acquire the capacity to perceive and comprehend language (in other words, gain the ability to be aware of language and to understand it), as well as to produce and use words and sentences to communicate. Language acquisition involves structures, rules and representation. The capacity to use language successfully requires one to acquire a range of tools including phonology, morphology, syntax, semantics, and an extensive vocabulary. Language can be vocalized as in speech, or manual as in sign. Human language capacity is represented in the brain. Even though human language capacity is finite, one can say and understand an infinite number of sentences, which is based on a syntactic principle called recursion. Evidence suggests that every individual has three recursive mechanisms that allow sentences to go indeterminately. These three mechanisms are: ''relativization'', ''complementation'' and ''coordination''. There are two ma ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Denotation
In linguistics and philosophy, the denotation of an expression is its literal meaning. For instance, the English word "warm" denotes the property of being warm. Denotation is contrasted with other aspects of meaning including connotation. For instance, the word "warm" may evoke calmness or cosiness, but these associations are not part of the word's denotation. Similarly, an expression's denotation is separate from pragmatic inferences it may trigger. For instance, describing something as "warm" often implicates that it is not hot, but this is once again not part of the word's denotation. Denotation plays a major role in several fields. Within philosophy of language, denotation is studied as an important aspect of meaning. In mathematics and computer science, assignments of denotations are assigned to expressions are a crucial step in defining interpreted formal languages. The main task of formal semantics is to reverse engineer the computational system which assigns denotations ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Lexicon
A lexicon is the vocabulary of a language or branch of knowledge (such as nautical or medical). In linguistics, a lexicon is a language's inventory of lexemes. The word ''lexicon'' derives from Koine Greek language, Greek word (), neuter of () meaning 'of or for words'. Linguistic theories generally regard human languages as consisting of two parts: a lexicon, essentially a catalogue of a language's words (its wordstock); and a grammar, a system of rules which allow for the combination of those words into meaningful sentences. The lexicon is also thought to include bound morphemes, which cannot stand alone as words (such as most affixes). In some analyses, compound words and certain classes of idiomatic expressions, collocations and other phrases are also considered to be part of the lexicon. Dictionary, Dictionaries are lists of the lexicon, in alphabetical order, of a given language; usually, however, bound morphemes are not included. Size and organization Items in the le ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |