Operator Grammar

Operator Grammar

Operator Grammar is a mathematical theory of human language that explains how language carries information. This theory is the culmination of the life work of Zellig Harris, with major publications toward the end of the last century. Operator Grammar proposes that each human language is a self-organizing system in which both the syntactic and semantic properties of a word are established purely in relation to other words. Thus, no external system (metalanguage) is required to define the rules of a language. Instead, these rules are learned through exposure to usage and through participation, as is the case with most social behavior. The theory is consistent with the idea that language evolved gradually, with each successive generation introducing new complexity and variation.

Operator Grammar posits three universal constraints: Dependency (certain words depend on the presence of other words to form an utterance), Likelihood (some combinations of words and their dependents are more likely than others) and Reduction (words in high likelihood combinations can be reduced to shorter forms, and sometimes omitted completely). Together these provide a theory of language information: dependency builds a predicate-argument structure; likelihood creates distinct meanings; reduction allows compact forms for communication.

Contents

Dependency

The fundamental mechanism of Operator Grammar is the dependency constraint: certain words (operators) require that one or more words (arguments) be present in an utterance. In the sentence John wears boots, the operator wears requires the presence of two arguments, such as John and boots. (This definition of dependency differs from other dependency grammars in which the arguments are said to depend on the operators.)

In each language the dependency relation among words gives rise to syntactic categories in which the allowable arguments of an operator are defined in terms of their dependency requirements. Class N contains words (e.g. John, boots) that do not require the presence of other words. Class ON contains the words (e.g. sleeps) that require exactly one word of type N. Class ONN contains the words (e.g. wears) that require two words of type N. Class OOO contains the words (e.g. because) that require two words of type O, as in John stumbles because John wears boots. Other classes include OO (is possible), ONNN (put), OON (with, surprise), ONO (know), ONNO (ask) and ONOO (attribute).

The categories in Operator Grammar are universal and are defined purely in terms of how words relate to other words, and do not rely on an external set of categories such as noun, verb, adjective, adverb, preposition, conjunction, etc. The dependency properties of each word are observable through usage and therefore learnable.

Likelihood

The dependency constraint creates a structure (syntax) in which any word of the appropriate class can be an argument for a given operator. The likelihood constraint places additional restrictions on this structure by making some operator/argument combinations more likely than others. Thus, John wears hats is more likely than John wears snow which in turn is more likely than John wears vacation. The likelihood constraint creates meaning (semantics) by defining each word in terms of the words it can take as arguments, or of which it can be an argument.

Each word has a unique set of words with which it has been observed to occur called its selection. The coherent selection of a word is the set of words for which the dependency relation has above average likelihood. Words that are similar in meaning have similar coherent selection. This approach to meaning is self-organizing in that no external system is necessary to define what words mean. Instead, the meaning of the word is determined by its usage within a population of speakers. Patterns of frequent use are observable and therefore learnable. New words can be introduced at any time and defined through usage.

Reduction

The reduction constraint acts on high likelihood combinations of operators and arguments and makes more compact forms. Certain reductions allow words to be omitted completely from an utterance. For example, I expect John to come is reducible to I expect John, because to come is highly likely under expect. The sentence John wears boots and John wears hats can be reduced to John wears boots and hats because repetition of the first argument John under the operator and is highly likely. John reads things can be reduced to John reads, because the argument things has high likelihood of occurring under any operator.

Certain reductions reduce words to shorter forms, creating pronouns, suffixes and prefixes (morphology). John wears boots and John wears hats can be reduced to John wears boots and he wears hats, where the pronoun he is a reduced form of John. Suffixes and prefixes can be obtained by appending other freely occurring words, or variants of these. John is able to be liked can be reduced to John is likeable. John is thoughtful is reduced from John is full of thought, and John is anti-war from John is against war.

Modifiers are the result of several of these kinds of reductions, which give rise to adjectives, adverbs, prepositional phrases, subordinate clauses, etc.

  1. John wears boots; the boots are of leather (two sentences joined by semicolon operator) →
  2. John wears boots which are of leather (reduction of repeated noun to relative pronoun) →
  3. John wears boots of leather (omission of high likelihood phrase which are) →
  4. John wears leather boots (omission of high likelihood operator of, transposition of short modifier to left of noun)

Each language has a unique set of reductions. For example, some languages have morphology and some don’t; some transpose short modifiers and some do not. Each word in a language participates only in certain kinds of reductions. However, in each case, the reduced material can be reconstructed from knowledge of what is likely in the given operator/argument combination. The reductions in which each word participates are observable and therefore learnable, just as one learns a word’s dependency and likelihood properties.

Information

The importance of reductions in Operator Grammar is that they separate sentences that contain reduced forms from those that don’t (base sentences). All reductions are paraphrases, since they do not remove any information, just make sentences more compact. Thus, the base sentences contain all the information of the language and the reduced sentences are variants of these. Base sentences are made up of simple words without modifiers and largely without affixes, e.g. Snow falls, Sheep eat grass, John knows sheep eat grass, That sheep eat snow surprises John.

Each operator in a sentence makes a contribution in information according to its likelihood of occurrence with its arguments. Highly expected combinations have low information; rare combinations have high information. The precise contribution of an operator is determined by its selection, the set of words with which it occurs with high frequency. The arguments boots, hats, sheep, grass and snow differ in meaning according to the operators for which they can appear with high likelihood in first or second argument position. For example, snow is expected as first argument of fall but not of eat, while the reverse is true of sheep. Similarly, the operators eat, devour, chew and swallow differ in meaning to the extent that the arguments they select and the operators that select them differ.

Operator Grammar predicts that the information carried by a sentence is the accumulation of contributions of each argument and operator. The increment of information that a given word adds to a new sentences is determined by how it was used before. In turn, new usages stretch or even alter the information content associated with a word. Because this process is based on high frequency usage, the meanings of words are relatively stable over time, but can change in accordance with the needs of a linguistic community.

Bibliography

  • Harris, Zellig (1982), A Grammar of English on Mathematical Principles, New York: John Wiley and Sons, ISBN 0471029580 
  • Harris, Zellig (1988), Language and Information, New York: Columbia University Press, ISBN 0-231-06662-7 
  • Harris, Zellig (1989), The Form of Information in Science: Analysis of an immunology sublanguage, Springer, ISBN 90-277-2516-0 
  • Harris, Zellig (1991), A Theory of Language and Information: A Mathematical Approach, Oxford University Press, USA, ISBN 0-19-824224-7 

Wikimedia Foundation. 2010.

Игры ⚽ Нужна курсовая?

Look at other dictionaries:

  • Operator — may refer to: Contents 1 Music 2 Computers 3 Science and mathematics …   Wikipedia

  • Operator-precedence parser — An operator precedence parser is a bottom up parser that interprets an operator precedence grammar. For example, most calculators use operator precedence parsers to convert from the human readable infix notation with order of operations format… …   Wikipedia

  • Grammar framework — In theoretical linguistics, the following fundamental approaches towards constructing grammar frameworks for natural languages are distinguished:*Generative grammar: algorithmic (phrase structure grammars) **Transformational grammar (1960s)… …   Wikipedia

  • Operator-precedence grammar — An operator precedence grammar is a kind of grammar for formal languages. Technically, an operator precedence grammar is a context free grammar that has the property (among others[1]) that no production has either an empty right hand side or two… …   Wikipedia

  • Grammar induction — Grammatical induction, also known as grammatical inference or syntactic pattern recognition, refers to the process in machine learning of inducing a formal grammar (usually in the form of re write rules or productions ) from a set of observations …   Wikipedia

  • Operator (linguistics) — In linguistics, an operator is a special variety of determiner including the visible interrogatives, the quantifiers, and the hypothetical invisible pronoun denoted Op. Operators are differentiated from other determiners by their ability to… …   Wikipedia

  • operator — /ˈɒpəreɪtə / (say opuhraytuh) noun 1. a worker; one employed or skilled in operating a machine, apparatus, or the like: a wireless operator; telephone operator. 2. someone who conducts some working or industrial establishment, enterprise, or… …  

  • Dependency grammar — Hybrid constituency/dependency tree from the Quranic Arabic Corpus Dependency grammar (DG) is a class of syntactic theories developed by Lucien Tesnière. It is distinct from phrase structure grammars, as it lacks phrasal nodes. Structure is… …   Wikipedia

  • Parsing expression grammar — A parsing expression grammar, or PEG, is a type of analytic formal grammar that describes a formal language in terms of a set of rules for recognizing strings in the language. A parsing expression grammar essentially represents a recursive… …   Wikipedia

  • Functional Discourse Grammar — Die Functional Discourse Grammar (FDG) ist eine neuere Grammatiktheorie. Von führenden Vertretern der Functional Grammar (FG) wurde sie im ersten Jahrzehnt des 3. Jahrtausends vor allem in Amsterdam in Weiterentwicklung der FG entwickelt. Über… …   Deutsch Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”