next up previous contents
Next: Constraint Logic Programming Up: Linguistic and Formal Foundations Previous: Linguistic and Formal Foundations

Constraint-based Grammars

 

Since the last decade a family of linguistic theories known under the term constraint-based grammar theories play an important role within the field of natural language processing. One of the reasons for their importance is that they are declarative, i.e., are able to express grammatical knowledge without regard to any specific use. Since they also integrate information from different levels of linguistic description uniformly they are very well suited as the linguistic base of a reversible system. In this sense, constraint-based grammars are reversible.

The common design feature of these formalisms is that they model linguistic entities (e.g., words and phrases) `` tex2html_wrap_inline10615 as partial information structures, which mutually constrain possible collocations of phonological structure, syntactic structure, semantic content, and contextual factors in actual linguistic situations. Such objects are in essence data structures which specify values for attributes; their capability to bear information of non-trivial complexity arises from their potential for recursive embedding (the value of some attribute may be another information structure with internal structure of its own) and structure-sharing (one and and the same structure may occur as the value of distinct attributes in a larger structure)'' [Pollard and Sag1987], page 7. Since, these informational structures are viewed as partial, they are used to represent only necessary restrictions or constraints of such properties a linguistic object must fulfill; hence the name constraint-based theories. To give an example, consider the following simplified attribute-value representation (also known as feature structure) in the commonly used feature-matrix notation, which is intended to represent the (partial) information about the utterance ``peter cries'':

displaymath10645

This structure expresses that the utterance ``peter cries'' is a linguistic object of class sentence, whose phonological representation is peter,cries, which has been formed from two objects, one for the utterance ``peter'' and one for ``cries'' and that the semantic content is the same as that of ``cries'' which by itself is built by combining the semantic structure of `cry' and `the-peter'. Clearly, this data structure is only a model of the `real' linguistic object. However, it characterizes relevant properties of the information the `real' object conveys.

The underlying assumption of all constraint-based theories is that in the actual utterance situation the specification of linguistic objects is built in a monotonic way independently of the specific processing task at hand (e.g., parsing, generation) from different sources, including lexicon and grammar rules, as well as from language universal and language specific principles. Building larger objects from smaller ones is assumed to be performed by some sort of constraint-solving whose task is to collect all constraints of the various attending objects, so that the resulting structure describes a consistent, i.e., grammatical linguistic object.gif

Most important from a reversibility standpoint is that the theories only characterize what constraints are important during natural language use, not in what order they are applied. Thus they are purely declarative. Furthermore, since almost all theories assume that a natural language grammar not only describes the correct sentences of a language but also the semantic structure of grammatically well-formed sentences, they are perfectly well suited to a reversible system, because they are neutral with respect to interpretation and production.

In the last few years constraint-based formalisms have undergone a rigorous formal investigation (consider for example [Shieber1989, Smolka1988, Smolka1992]). This has led to a general characterization of constraint-based formalisms where feature structures are considered to constitute a semantic domain and constraints are considered syntactic representations of such `semantic structures'. This logical view has several advantages. On the one hand, it has been possible to properly incorporate concepts like disjunction or negation as part of the (syntactic) constraint language and to interpret them relative to a given domain of feature structures (usually defined as graph-like or tree-like structures). On the other hand it has been possible to combine constraint-based formalisms with logic programming, which fits into a new research area known under the term constraint logic programming (CLP) [Jaffar and Lassez1987]. This enables us to extend the view of natural language as deduction for essentially arbitrary constraint-based formalisms.

A general characterization of constraint logic programming is given in [Höhfeld and Smolka1988]. They show that the nice properties of logic programs extend to definite clause specifications over arbitrary constraint languages.

We will use the scheme of [Höhfeld and Smolka1988] as our underlying formal language. We therefore first summarize the most important definitions and statements given there. We then specify a concrete constraint language also called which is based on the definitions of [Smolka1992] and [VanNoord1993]. This constraint language will be used as the central data structure for representing linguistic objects. The relational extensions provided by then serve as the basis for representing grammatical rules, i.e., complex compositional entities. The use of the presented formalism in writing grammars is illustrated in section 3.3.


next up previous contents
Next: Constraint Logic Programming Up: Linguistic and Formal Foundations Previous: Linguistic and Formal Foundations

Guenter Neumann
Mon Oct 5 14:01:36 MET DST 1998