Going on 5 years ago, I wrote part 1. Now, finally, it’s time for the rest of the story.
Continue reading “Confessions of a production rule vendor (part 2)”
systems that know and understand and think and learn
Going on 5 years ago, I wrote part 1. Now, finally, it’s time for the rest of the story.
Continue reading “Confessions of a production rule vendor (part 2)”
In preparing for some natural language generation[1], I came across some work on natural logic[2][3] and reasoning by textual entailment[4] (RTE) by Richard Bergmair in his PhD at Cambridge:
The work he describes overlaps our approach to robust inference from the deep, variable-precision semantics that result from linguistic analysis and disambiguation using the English Resource Grammar (ERG) and the Linguist™.
We’re working with the English Resource Grammar (ERG), OWL, and Vulcan’s SILK to educate the machine by translating textbooks into defeasible logic. Part of this involves an ontology that models semantics more deeply than the ERG, which is based on head-driven phrase structure grammar (HPSG), which provides deeper parsing and, with the ERG and the DELPH-IN infrastructure, also provides a simple under-specified semantic representation called minimal recursion semantics (MRS).
We’re having a great time using OWL to clarify and enrich the semantics of the rich model underlying the ERG. Here’s an example, FYI. If you’d like to know more (or help), please drop us a line! Overall the project will demonstrate our capabilities for transforming everyday sentences into RIF and business rule languages using SBVR extended with defeasibility and other capabilities, all modeled in the same OWL ontology.
What triggered this blog entry was a bit of a surprise in seeing that whether or not an adjective could be used depictively is sometimes encoded in the lexicon. This is one of the problems of TDL versus a description-logic based model with more expressiveness. It results in more lexical entries than necessary, which has been discussed by others when contrasted with the attributed logic engine (ALE), for example.
In trying to model the semantics of words like ‘same’ and ‘different’, we are scratching our heads about these lines from the ERG’s lexicon:
One of the interesting things about lexicalized grammars is that lexical entries (i.e., ‘words’) are described with almost arbitrary combinations of their lexical, syntactic, and semantic characteristics.
The preceding code is expressed in a type description language (TDL) used by the Lisp-based LKB (and its C++ counterpart, PET, which are unification-based parsers that produce a chart of plausible parses with some efficiency. What is given above is already deeper than what you can expect from a statistical parser (but richer descriptions of lexical entries promises to make statistical parsing much better, too).
Unfortunately, there is no available documentation on why the ERG was designed as it is, so the meaning of the above is difficult to interpret. For example, the types of lexical entries (the symbols ending in ‘_le’) referenced above are defined as follows:
Needless to say, that’s a mouthful! Chasing this down, the following ‘informs’ us that “the same”, which uses type #2 above, is defined using the following lexical types:
Which is to say that it is non-conjunctive, complements a head to form a phrase, can’t be affixed, cannot constitute a main clause, and is a word.
The fact that the lexical entry for “the same” is adjectival is given the definition of the following type(s) used in the SYNSEM feature:
Which is to say that it is a comparative or superlative adjectival word (even though it consists of two lexemes in its ‘orthography’) that involves two semantic arguments including one complement which may be unexpressed prepositional phrase. A comparative or superlative adjective, in turn, is non-conjunctive, complements a head to form a phrase, is non-affix bearing (?), and non-clausal, as defined by the type ‘nonc-hm-nab’ above.
The types used in the syntax and semantic (i.e., SYNSEM) feature of the two lexical types are defined as follows (none of which is documented):
In a moment, we’ll discuss the types used in the second of these, but first, some basics on the semantics that are mixed with the syntax above.
In effect, the above indicates that a new ‘elementary predication’ will be needed in the MRS to represent the adjectival relationship in the logic derived in the course of parsing (i.e., that’s what ‘unsp_ind’ means, although it’s not documented, which I will try not to bemoan much further.)
The following indicates that the newly formed elementary predicate is not (initially) within any scope and that it has two arguments whose semantics (i.e., their RELations) are concatenated for propagation into the list of elementary predications that will constitute the MRS for any parses found.
The last of these expresses that the constuction is lexical rather than phrasal (which includes clausal in the ERG).
Continuing with the definition of “the same” as an adjective, the following finally clarifies what it means to be a basic adjective:
Well, ‘clarifies’ might not have been the right word! Essentially, it indicates that the adjective may have an optional degree specifier (which semantically modifies the predicate of the adjective) and that the predicate specified in the lexical entry becomes the predicate used in the MRS. The rest is defined below:
Which ends with a bunch of basic setup types except for constraining for relation for an adjective to be ‘basically adjectivally’ on the first two lines. Also on these first two lines, it specifies that its subject and its specifier, if any, must be completed (i.e., empty) and agree with its non-conjunctive argument (which is not to say that it cannot be conjunctive, but that it modifies the conjunction as a whole, if so.) Whether or not it is expressed will determine if there are any further predicates about its arguments or if its unexpressed argument is identified by an otherwise unreferenced variable in any resulting MRS.
The lexical grounding of this type specification is given below, indicating that it may (or not) have phonology (e.g., pronunciation, such as whether its onset is voiced) and if and how and with what punctuation it may appear, if any. In general, a semantic argument may be lexical or phrasal and optional but if it appears it corresponds to some semantic index (think variable) in sort of predicate in any resulting MRS. (The *_min types do not constrain the values of their features any further).
Note that an adjective is not possess-able and that it modifies something nominal (or a title) and that if it has a specifier that it is a quantifier or degree (e.g., ‘very’). Again, an adjective cannot function as a main clause or be conjunctive (in and of itself).
Finally, if you look far above you will see that the basic semantics of an adjective with an additional semantic argument is ‘intersective’, as in:
Here, the length 0 difference list and the following definitions indicate that intersective semantics do not accept anything but local modification:
AVM stands for ‘attribute value matrix’, which is the structure by which types and their features are defined (with nesting and unification constraints using # to indicate equality).
By now you’re probably getting the idea that there is fairly significant model of the English language, including its lexical and syntactic aspects, but if you look there is a lot about semantics here, too.